In this paper, we explore how to utilize pre-trained language model to perform few-shot text classification where only a few annotated examples are given for each class. Since using traditional cross-entropy loss to fine-tune language model under this scenario causes serious overfitting and leads to sub-optimal generalization of model, we adopt supervised contrastive learning on few labeled data and consistency-regularization on vast unlabeled data. Moreover, we propose a novel contrastive consistency to further boost model performance and refine sentence representation. After conducting extensive experiments on four datasets, we demonstrate that our model (FTCC) can outperform state-of-the-art methods and has better robustness.Comment: 8 p...
Domain-specific text classification faces the challenge of scarce labeled data due to the high cost ...
Some NLP tasks can be solved in a fully unsupervised fashion by providing a pretrained language mode...
International audienceDuring the last few years, deep supervised learning models have been shown to ...
Few-shot text classification has recently been promoted by the meta-learning paradigm which aims to ...
The impressive performance of GPT-3 using natural language prompts and in-context learning has inspi...
A two-stage training paradigm consisting of sequential pre-training and meta-training stages has bee...
In recent years, the community of natural language processing (NLP) has seen amazing progress in the...
In recent years, there has been significant progress in developing pre-trained language models for N...
Few-shot learning aims to train models that can be generalized to novel classes with only a few samp...
Traditional text classification requires thousands of annotated data or an additional Neural Machine...
Large-scale multi-label text classification (LMTC) aims to associate a document with its relevant la...
Humans can obtain the knowledge of novel visual concepts from language descriptions, and we thus use...
Data augmentation techniques are widely used for enhancing the performance of machine learning model...
Data augmentation has been an important ingredient for boosting performances of learned models. Prio...
Prompt learning recently become an effective linguistic tool to motivate the PLMs' knowledge on few-...
Domain-specific text classification faces the challenge of scarce labeled data due to the high cost ...
Some NLP tasks can be solved in a fully unsupervised fashion by providing a pretrained language mode...
International audienceDuring the last few years, deep supervised learning models have been shown to ...
Few-shot text classification has recently been promoted by the meta-learning paradigm which aims to ...
The impressive performance of GPT-3 using natural language prompts and in-context learning has inspi...
A two-stage training paradigm consisting of sequential pre-training and meta-training stages has bee...
In recent years, the community of natural language processing (NLP) has seen amazing progress in the...
In recent years, there has been significant progress in developing pre-trained language models for N...
Few-shot learning aims to train models that can be generalized to novel classes with only a few samp...
Traditional text classification requires thousands of annotated data or an additional Neural Machine...
Large-scale multi-label text classification (LMTC) aims to associate a document with its relevant la...
Humans can obtain the knowledge of novel visual concepts from language descriptions, and we thus use...
Data augmentation techniques are widely used for enhancing the performance of machine learning model...
Data augmentation has been an important ingredient for boosting performances of learned models. Prio...
Prompt learning recently become an effective linguistic tool to motivate the PLMs' knowledge on few-...
Domain-specific text classification faces the challenge of scarce labeled data due to the high cost ...
Some NLP tasks can be solved in a fully unsupervised fashion by providing a pretrained language mode...
International audienceDuring the last few years, deep supervised learning models have been shown to ...