The impressive performance of GPT-3 using natural language prompts and in-context learning has inspired work on better fine-tuning of moderately-sized models under this paradigm. Following this line of work, we present a contrastive learning framework that clusters inputs from the same class for better generality of models trained with only limited examples. Specifically, we propose a supervised contrastive framework that clusters inputs from the same class under different augmented "views" and repel the ones from different classes. We create different "views" of an example by appending it with different language prompts and contextual demonstrations. Combining a contrastive loss with the standard masked language modeling (MLM) loss in prom...
Recent advances on large pre-trained language models (PLMs) lead impressive gains on natural languag...
Recent few-shot methods, such as parameter-efficient fine-tuning (PEFT) and pattern exploiting train...
International audienceLarge language models have recently been shown to attain reasonable zero-shot ...
Prompt learning recently become an effective linguistic tool to motivate the PLMs' knowledge on few-...
In recent years, there has been significant progress in developing pre-trained language models for N...
In this paper, we explore how to utilize pre-trained language model to perform few-shot text classif...
Few-shot text classification has recently been promoted by the meta-learning paradigm which aims to ...
Pretrained language models can be effectively stimulated by textual prompts or demonstrations, espec...
A two-stage training paradigm consisting of sequential pre-training and meta-training stages has bee...
Pretraining deep neural networks to perform language modeling - that is, to reconstruct missing word...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
Large-scale pre-trained language models have contributed significantly to natural language processin...
Few-shot learning aims to train models that can be generalized to novel classes with only a few samp...
Prompt-based learning has shown considerable promise in reformulating various downstream tasks as cl...
Deep learning has recently driven remarkable progress in several applications, including image class...
Recent advances on large pre-trained language models (PLMs) lead impressive gains on natural languag...
Recent few-shot methods, such as parameter-efficient fine-tuning (PEFT) and pattern exploiting train...
International audienceLarge language models have recently been shown to attain reasonable zero-shot ...
Prompt learning recently become an effective linguistic tool to motivate the PLMs' knowledge on few-...
In recent years, there has been significant progress in developing pre-trained language models for N...
In this paper, we explore how to utilize pre-trained language model to perform few-shot text classif...
Few-shot text classification has recently been promoted by the meta-learning paradigm which aims to ...
Pretrained language models can be effectively stimulated by textual prompts or demonstrations, espec...
A two-stage training paradigm consisting of sequential pre-training and meta-training stages has bee...
Pretraining deep neural networks to perform language modeling - that is, to reconstruct missing word...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
Large-scale pre-trained language models have contributed significantly to natural language processin...
Few-shot learning aims to train models that can be generalized to novel classes with only a few samp...
Prompt-based learning has shown considerable promise in reformulating various downstream tasks as cl...
Deep learning has recently driven remarkable progress in several applications, including image class...
Recent advances on large pre-trained language models (PLMs) lead impressive gains on natural languag...
Recent few-shot methods, such as parameter-efficient fine-tuning (PEFT) and pattern exploiting train...
International audienceLarge language models have recently been shown to attain reasonable zero-shot ...