Few-shot in-context learning (ICL) enables pre-trained language models to perform a previously-unseen task without any gradient-based training by feeding a small number of training examples as part of the input. ICL incurs substantial computational, memory, and storage costs because it involves processing all of the training examples every time a prediction is made. Parameter-efficient fine-tuning (e.g. adapter modules, prompt tuning, sparse update methods, etc.) offers an alternative paradigm where a small set of parameters are trained to enable a model to perform the new task. In this paper, we rigorously compare few-shot ICL and parameter-efficient fine-tuning and demonstrate that the latter offers better accuracy as well as dramatically...
Self-supervised pretraining has made few-shot learning possible for many NLP tasks. But the pretrain...
When scaled to hundreds of billions of parameters, pretrained language models such as GPT-3 (Brown e...
Domain-specific text classification faces the challenge of scarce labeled data due to the high cost ...
We present a new method LiST is short for Lite Prompted Self-Training for parameter-efficient fine-t...
Through in-context learning (ICL), large-scale language models are effective few-shot learners witho...
Pretrained large language models (LLMs) are strong in-context learners that are able to perform few-...
Recent few-shot methods, such as parameter-efficient fine-tuning (PEFT) and pattern exploiting train...
Business analytics and machine learning have become essential success factors for various industries...
Despite achieving state-of-the-art zero-shot performance, existing vision-language models still fall...
Business analytics and machine learning have become essential success factors for various industries...
Few-shot Learning (FSL) is aimed to make predictions based on a limited number of samples. Structure...
Language models (LMs) with less than 100B parameters are known to perform poorly on chain-of-thought...
Multi-task learning (MTL), instruction tuning, and prompting have recently been shown to improve the...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
Recently, there has been an increasing interest in models that generate natural language explanation...
Self-supervised pretraining has made few-shot learning possible for many NLP tasks. But the pretrain...
When scaled to hundreds of billions of parameters, pretrained language models such as GPT-3 (Brown e...
Domain-specific text classification faces the challenge of scarce labeled data due to the high cost ...
We present a new method LiST is short for Lite Prompted Self-Training for parameter-efficient fine-t...
Through in-context learning (ICL), large-scale language models are effective few-shot learners witho...
Pretrained large language models (LLMs) are strong in-context learners that are able to perform few-...
Recent few-shot methods, such as parameter-efficient fine-tuning (PEFT) and pattern exploiting train...
Business analytics and machine learning have become essential success factors for various industries...
Despite achieving state-of-the-art zero-shot performance, existing vision-language models still fall...
Business analytics and machine learning have become essential success factors for various industries...
Few-shot Learning (FSL) is aimed to make predictions based on a limited number of samples. Structure...
Language models (LMs) with less than 100B parameters are known to perform poorly on chain-of-thought...
Multi-task learning (MTL), instruction tuning, and prompting have recently been shown to improve the...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
Recently, there has been an increasing interest in models that generate natural language explanation...
Self-supervised pretraining has made few-shot learning possible for many NLP tasks. But the pretrain...
When scaled to hundreds of billions of parameters, pretrained language models such as GPT-3 (Brown e...
Domain-specific text classification faces the challenge of scarce labeled data due to the high cost ...