Nowadays, owing to the superior capacity of the large pre-trained language models (PLM), the PLM-based zero-shot learning has shown promising performances on various natural language processing tasks. There are emerging interests in further exploring the zero-shot learning potential of PLMs. Among them, ZeroGen attempts to purely use PLM to generate data and train a tiny model without relying on any task-specific annotation. Despite its remarkable results, we observe that the synthesized data from PLM contains a significant portion of samples with low quality, overfitting on such data greatly hampers the performance of the trained model and makes it unreliable for deployment.Since no gold data is accessible in zero-shot scenario, it is hard...
The reusability of state-of-the-art Pre-trained Language Models (PLMs) is often limited by their gen...
Large pre-trained models such as CLIP or ALIGN offer consistent accuracy across a range of data dist...
Pretrained language models have shown success in various areas of natural language processing, inclu...
There is a growing interest in dataset generation recently due to the superior generative capacity o...
Pretrained language models (PLMs) have demonstrated remarkable performance in various natural langua...
Traditional text classification approaches often require a good amount of labeled data, which is dif...
We propose a multitask pretraining approach ZeroPrompt for zero-shot generalization, focusing on tas...
In recent years, the community of natural language processing (NLP) has seen amazing progress in the...
One of the most impressive results of recent NLP history is the ability of pre-trained language mode...
Existing solutions to zero-shot text classification either conduct prompting with pre-trained langu...
This paper studies the use of language models as a source of synthetic unlabeled text for NLP. We fo...
International audienceLarge language models have recently been shown to attain reasonable zero-shot ...
We present a new method LiST is short for Lite Prompted Self-Training for parameter-efficient fine-t...
High-quality instruction-tuning data is critical to improving LLM capabilities. Existing data collec...
Black-Box Tuning (BBT) is a derivative-free approach to optimize continuous prompt tokens prepended ...
The reusability of state-of-the-art Pre-trained Language Models (PLMs) is often limited by their gen...
Large pre-trained models such as CLIP or ALIGN offer consistent accuracy across a range of data dist...
Pretrained language models have shown success in various areas of natural language processing, inclu...
There is a growing interest in dataset generation recently due to the superior generative capacity o...
Pretrained language models (PLMs) have demonstrated remarkable performance in various natural langua...
Traditional text classification approaches often require a good amount of labeled data, which is dif...
We propose a multitask pretraining approach ZeroPrompt for zero-shot generalization, focusing on tas...
In recent years, the community of natural language processing (NLP) has seen amazing progress in the...
One of the most impressive results of recent NLP history is the ability of pre-trained language mode...
Existing solutions to zero-shot text classification either conduct prompting with pre-trained langu...
This paper studies the use of language models as a source of synthetic unlabeled text for NLP. We fo...
International audienceLarge language models have recently been shown to attain reasonable zero-shot ...
We present a new method LiST is short for Lite Prompted Self-Training for parameter-efficient fine-t...
High-quality instruction-tuning data is critical to improving LLM capabilities. Existing data collec...
Black-Box Tuning (BBT) is a derivative-free approach to optimize continuous prompt tokens prepended ...
The reusability of state-of-the-art Pre-trained Language Models (PLMs) is often limited by their gen...
Large pre-trained models such as CLIP or ALIGN offer consistent accuracy across a range of data dist...
Pretrained language models have shown success in various areas of natural language processing, inclu...