We present a new method LiST is short for Lite Prompted Self-Training for parameter-efficient fine-tuning of large pre-trained language models (PLMs) for few-shot learning. LiST improves over recent methods that adopt prompt-based fine-tuning (FN) using two key techniques. The first is the use of self-training to leverage large amounts of unlabeled data for prompt-based FN in few-shot settings. We use self-training in conjunction with meta-learning for re-weighting noisy pseudo-prompt labels. Self-training is expensive as it requires updating all the model parameters repetitively. Therefore, we use a second technique for light-weight fine-tuning where we introduce a small number of task-specific parameters that are fine-tuned during self-tr...
Why can pre-trained language models (PLMs) learn universal representations and effectively adapt to ...
Pretrained large language models (LLMs) are strong in-context learners that are able to perform few-...
Few-shot Learning (FSL) is aimed to make predictions based on a limited number of samples. Structure...
Few-shot in-context learning (ICL) enables pre-trained language models to perform a previously-unsee...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
Deploying large language models (LLMs) is challenging because they are memory inefficient and comput...
Language models (LMs) with less than 100B parameters are known to perform poorly on chain-of-thought...
Recent advances on large pre-trained language models (PLMs) lead impressive gains on natural languag...
Domain-specific text classification faces the challenge of scarce labeled data due to the high cost ...
In recent years, there has been significant progress in developing pre-trained language models for N...
Recently, there has been an increasing interest in models that generate natural language explanation...
Prompt-tuning has shown appealing performance in few-shot classification by virtue of its capability...
Recent few-shot methods, such as parameter-efficient fine-tuning (PEFT) and pattern exploiting train...
Very large language models (LLMs), such as GPT-3 and Codex have achieved state-of-the-art performanc...
Adopting a two-stage paradigm of pretraining followed by fine-tuning, Pretrained Language Models (PL...
Why can pre-trained language models (PLMs) learn universal representations and effectively adapt to ...
Pretrained large language models (LLMs) are strong in-context learners that are able to perform few-...
Few-shot Learning (FSL) is aimed to make predictions based on a limited number of samples. Structure...
Few-shot in-context learning (ICL) enables pre-trained language models to perform a previously-unsee...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
Deploying large language models (LLMs) is challenging because they are memory inefficient and comput...
Language models (LMs) with less than 100B parameters are known to perform poorly on chain-of-thought...
Recent advances on large pre-trained language models (PLMs) lead impressive gains on natural languag...
Domain-specific text classification faces the challenge of scarce labeled data due to the high cost ...
In recent years, there has been significant progress in developing pre-trained language models for N...
Recently, there has been an increasing interest in models that generate natural language explanation...
Prompt-tuning has shown appealing performance in few-shot classification by virtue of its capability...
Recent few-shot methods, such as parameter-efficient fine-tuning (PEFT) and pattern exploiting train...
Very large language models (LLMs), such as GPT-3 and Codex have achieved state-of-the-art performanc...
Adopting a two-stage paradigm of pretraining followed by fine-tuning, Pretrained Language Models (PL...
Why can pre-trained language models (PLMs) learn universal representations and effectively adapt to ...
Pretrained large language models (LLMs) are strong in-context learners that are able to perform few-...
Few-shot Learning (FSL) is aimed to make predictions based on a limited number of samples. Structure...