Recent few-shot methods, such as parameter-efficient fine-tuning (PEFT) and pattern exploiting training (PET), have achieved impressive results in label-scarce settings. However, they are difficult to employ since they are subject to high variability from manually crafted prompts, and typically require billion-parameter language models to achieve high accuracy. To address these shortcomings, we propose SetFit (Sentence Transformer Fine-tuning), an efficient and prompt-free framework for few-shot fine-tuning of Sentence Transformers (ST). SetFit works by first fine-tuning a pretrained ST on a small number of text pairs, in a contrastive Siamese manner. The resulting model is then used to generate rich text embeddings, which are used to train...
Recent prompt-based approaches allow pretrained language models to achieve strong performances on fe...
In real-world applications of natural language generation, target sentences are often required to sa...
When scaled to hundreds of billions of parameters, pretrained language models such as GPT-3 (Brown e...
Recent few-shot learning methods, such as parameter-efficient fine-tuning (PEFT) and pattern exploit...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
Domain-specific text classification faces the challenge of scarce labeled data due to the high cost ...
Few-shot in-context learning (ICL) enables pre-trained language models to perform a previously-unsee...
We present a new method LiST is short for Lite Prompted Self-Training for parameter-efficient fine-t...
Providing pretrained language models with simple task descriptions in natural language enables them ...
Providing pretrained language models with simple task descriptions in natural language enables them ...
Providing pretrained language models with simple task descriptions in natural language enables them ...
The impressive performance of GPT-3 using natural language prompts and in-context learning has inspi...
Providing pretrained language models with simple task descriptions in natural language enables them ...
Recent advances on large pre-trained language models (PLMs) lead impressive gains on natural languag...
When scaled to hundreds of billions of parameters, pretrained language models such as GPT-3 (Brown e...
Recent prompt-based approaches allow pretrained language models to achieve strong performances on fe...
In real-world applications of natural language generation, target sentences are often required to sa...
When scaled to hundreds of billions of parameters, pretrained language models such as GPT-3 (Brown e...
Recent few-shot learning methods, such as parameter-efficient fine-tuning (PEFT) and pattern exploit...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
Domain-specific text classification faces the challenge of scarce labeled data due to the high cost ...
Few-shot in-context learning (ICL) enables pre-trained language models to perform a previously-unsee...
We present a new method LiST is short for Lite Prompted Self-Training for parameter-efficient fine-t...
Providing pretrained language models with simple task descriptions in natural language enables them ...
Providing pretrained language models with simple task descriptions in natural language enables them ...
Providing pretrained language models with simple task descriptions in natural language enables them ...
The impressive performance of GPT-3 using natural language prompts and in-context learning has inspi...
Providing pretrained language models with simple task descriptions in natural language enables them ...
Recent advances on large pre-trained language models (PLMs) lead impressive gains on natural languag...
When scaled to hundreds of billions of parameters, pretrained language models such as GPT-3 (Brown e...
Recent prompt-based approaches allow pretrained language models to achieve strong performances on fe...
In real-world applications of natural language generation, target sentences are often required to sa...
When scaled to hundreds of billions of parameters, pretrained language models such as GPT-3 (Brown e...