Probing is a popular method to discern what linguistic information is contained in the representations of pre-trained language models. However, the mechanism of selecting the probe model has recently been subject to intense debate, as it is not clear if the probes are merely extracting information or modeling the linguistic property themselves. To address this challenge, this paper introduces a novel model-free approach to probing, by formulating probing as a prompting task. We conduct experiments on five probing tasks and show that our approach is comparable or better at extracting information than diagnostic probes while learning much less on its own. We further combine the probing via prompting approach with attention head pruning to ana...
Classifiers trained on auxiliary probing tasks are a popular tool to analyze the representations lea...
Knowledge probing is crucial for understanding the knowledge transfer mechanism behind the pre-train...
Through in-context learning (ICL), large-scale language models are effective few-shot learners witho...
Probing is a popular approach to understand what linguistic information is contained in the represen...
Analysing whether neural language models encode linguistic information has become popular in NLP. On...
Through in-context learning (ICL), large-scale language models are effective few-shot learners witho...
Over the past decades natural language processing has evolved from a niche research area into a fast...
There is an ongoing debate in the NLP community whether modern language models contain linguistic k...
Previous work on probing word representations for linguistic knowledge has focused on interpolation ...
Neural language models have drastically changed the landscape of natural language processing (NLP). ...
Title: Investigating Large Language Models' Representations Of Plurality Through Probing Interventio...
Pre-trained contextual representations have led to dramatic performance improvements on a range of d...
Thesis: Ph. D., Massachusetts Institute of Technology, Department of Brain and Cognitive Sciences, F...
The outstanding performance recently reached by Neural Language Models (NLMs) across many Natural La...
Comunicació presentada a la Conference on Empirical Methods in Natural Language Processing (EMNLP 20...
Classifiers trained on auxiliary probing tasks are a popular tool to analyze the representations lea...
Knowledge probing is crucial for understanding the knowledge transfer mechanism behind the pre-train...
Through in-context learning (ICL), large-scale language models are effective few-shot learners witho...
Probing is a popular approach to understand what linguistic information is contained in the represen...
Analysing whether neural language models encode linguistic information has become popular in NLP. On...
Through in-context learning (ICL), large-scale language models are effective few-shot learners witho...
Over the past decades natural language processing has evolved from a niche research area into a fast...
There is an ongoing debate in the NLP community whether modern language models contain linguistic k...
Previous work on probing word representations for linguistic knowledge has focused on interpolation ...
Neural language models have drastically changed the landscape of natural language processing (NLP). ...
Title: Investigating Large Language Models' Representations Of Plurality Through Probing Interventio...
Pre-trained contextual representations have led to dramatic performance improvements on a range of d...
Thesis: Ph. D., Massachusetts Institute of Technology, Department of Brain and Cognitive Sciences, F...
The outstanding performance recently reached by Neural Language Models (NLMs) across many Natural La...
Comunicació presentada a la Conference on Empirical Methods in Natural Language Processing (EMNLP 20...
Classifiers trained on auxiliary probing tasks are a popular tool to analyze the representations lea...
Knowledge probing is crucial for understanding the knowledge transfer mechanism behind the pre-train...
Through in-context learning (ICL), large-scale language models are effective few-shot learners witho...