Recent few-shot methods, such as parameter-efficient fine-tuning (PEFT) and pattern exploiting training (PET), have achieved impressive results in label-scarce settings. However, they are difficult to employ since they are subject to high variability from manually crafted prompts, and typically require billion-parameter language models to achieve high accuracy. To address these shortcomings, we propose SetFit (Sentence Transformer Fine-tuning), an efficient and prompt-free framework for few-shot fine-tuning of Sentence Transformers (ST). SetFit works by first fine-tuning a pretrained ST on a small number of text pairs, in a contrastive Siamese manner. The resulting model is then used to generate rich text embeddings, which are used to train...
Some NLP tasks can be solved in a fully unsupervised fashion by providing a pretrained language mode...
In this paper, we explore how to utilize pre-trained language model to perform few-shot text classif...
When primed with only a handful of training samples, very large, pretrained language models such as ...
Recent few-shot learning methods, such as parameter-efficient fine-tuning (PEFT) and pattern exploit...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
Domain-specific text classification faces the challenge of scarce labeled data due to the high cost ...
Few-shot in-context learning (ICL) enables pre-trained language models to perform a previously-unsee...
We present a new method LiST is short for Lite Prompted Self-Training for parameter-efficient fine-t...
Providing pretrained language models with simple task descriptions in natural language enables them ...
The impressive performance of GPT-3 using natural language prompts and in-context learning has inspi...
Recent advances on large pre-trained language models (PLMs) lead impressive gains on natural languag...
When scaled to hundreds of billions of parameters, pretrained language models such as GPT-3 (Brown e...
Recent prompt-based approaches allow pretrained language models to achieve strong performances on fe...
In real-world applications of natural language generation, target sentences are often required to sa...
Large-scale pre-trained language models have contributed significantly to natural language processin...
Some NLP tasks can be solved in a fully unsupervised fashion by providing a pretrained language mode...
In this paper, we explore how to utilize pre-trained language model to perform few-shot text classif...
When primed with only a handful of training samples, very large, pretrained language models such as ...
Recent few-shot learning methods, such as parameter-efficient fine-tuning (PEFT) and pattern exploit...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
Domain-specific text classification faces the challenge of scarce labeled data due to the high cost ...
Few-shot in-context learning (ICL) enables pre-trained language models to perform a previously-unsee...
We present a new method LiST is short for Lite Prompted Self-Training for parameter-efficient fine-t...
Providing pretrained language models with simple task descriptions in natural language enables them ...
The impressive performance of GPT-3 using natural language prompts and in-context learning has inspi...
Recent advances on large pre-trained language models (PLMs) lead impressive gains on natural languag...
When scaled to hundreds of billions of parameters, pretrained language models such as GPT-3 (Brown e...
Recent prompt-based approaches allow pretrained language models to achieve strong performances on fe...
In real-world applications of natural language generation, target sentences are often required to sa...
Large-scale pre-trained language models have contributed significantly to natural language processin...
Some NLP tasks can be solved in a fully unsupervised fashion by providing a pretrained language mode...
In this paper, we explore how to utilize pre-trained language model to perform few-shot text classif...
When primed with only a handful of training samples, very large, pretrained language models such as ...