Through in-context learning (ICL), large-scale language models are effective few-shot learners without additional model fine-tuning. However, the ICL performance does not scale well with the number of available training sample as it is limited by the inherent input length constraint of the underlying language model. Meanwhile, many studies have revealed that language models are also powerful feature extractors, allowing them to be utilized in a black-box manner and enabling the linear probing paradigm, where lightweight discriminators are trained on top of the pre-extracted input representations. This paper proposes prompt-augmented linear probing (PALP), a hybrid of linear probing and ICL, which leverages the best of both worlds. PALP inh...
Large pre-trained vision-language models like CLIP have shown great potential in learning representa...
The impressive performance of GPT-3 using natural language prompts and in-context learning has inspi...
Few-shot in-context learning (ICL) enables pre-trained language models to perform a previously-unsee...
Through in-context learning (ICL), large-scale language models are effective few-shot learners witho...
Probing is a popular method to discern what linguistic information is contained in the representatio...
Large-scale pre-trained language models have contributed significantly to natural language processin...
In-context learning is a recent paradigm in natural language understanding, where a large pre-traine...
International audienceThe increasingly widespread adoption of large language models has highlighted ...
Pretrained large language models (LLMs) are strong in-context learners that are able to perform few-...
International audienceLarge language models have recently been shown to attain reasonable zero-shot ...
This paper investigates very low resource language model pretraining, when less than 100 thousand se...
Recent advances on large pre-trained language models (PLMs) lead impressive gains on natural languag...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
Prompt learning is a new paradigm in the Natural Language Processing (NLP) field which has shown imp...
Previous work on probing word representations for linguistic knowledge has focused on interpolation ...
Large pre-trained vision-language models like CLIP have shown great potential in learning representa...
The impressive performance of GPT-3 using natural language prompts and in-context learning has inspi...
Few-shot in-context learning (ICL) enables pre-trained language models to perform a previously-unsee...
Through in-context learning (ICL), large-scale language models are effective few-shot learners witho...
Probing is a popular method to discern what linguistic information is contained in the representatio...
Large-scale pre-trained language models have contributed significantly to natural language processin...
In-context learning is a recent paradigm in natural language understanding, where a large pre-traine...
International audienceThe increasingly widespread adoption of large language models has highlighted ...
Pretrained large language models (LLMs) are strong in-context learners that are able to perform few-...
International audienceLarge language models have recently been shown to attain reasonable zero-shot ...
This paper investigates very low resource language model pretraining, when less than 100 thousand se...
Recent advances on large pre-trained language models (PLMs) lead impressive gains on natural languag...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
Prompt learning is a new paradigm in the Natural Language Processing (NLP) field which has shown imp...
Previous work on probing word representations for linguistic knowledge has focused on interpolation ...
Large pre-trained vision-language models like CLIP have shown great potential in learning representa...
The impressive performance of GPT-3 using natural language prompts and in-context learning has inspi...
Few-shot in-context learning (ICL) enables pre-trained language models to perform a previously-unsee...