There is a growing interest in dataset generation recently due to the superior generative capacity of large pre-trained language models (PLMs). In this paper, we study a flexible and efficient zero-short learning method, ZeroGen. Given a zero-shot task, we first generate a dataset from scratch using PLMs in an unsupervised manner. Then, we train a tiny task model (e.g., LSTM) under the supervision of the synthesized dataset. This approach allows highly efficient inference as the final task model only has orders of magnitude fewer parameters comparing to PLMs (e.g., GPT2-XL). Apart from being annotation-free and efficient, we argue that ZeroGen can also provide useful insights from the perspective of data-free model-agnostic knowledge distil...
We study the problem of generating a training-free task-dependent visual classifier from text descri...
Our research focuses on solving the zero-shot text classification problem in NLP, with a particular ...
We present a deep generative model for Zero-Shot Learning (ZSL). Unlike most existing methods for th...
Nowadays, owing to the superior capacity of the large pre-trained language models (PLM), the PLM-bas...
Pretrained language models (PLMs) have demonstrated remarkable performance in various natural langua...
In recent years, the community of natural language processing (NLP) has seen amazing progress in the...
International audienceLarge language models have recently been shown to attain reasonable zero-shot ...
Traditional text classification approaches often require a good amount of labeled data, which is dif...
One of the most impressive results of recent NLP history is the ability of pre-trained language mode...
We propose a novel framework for zero-shot learning of topic-dependent language models, which enable...
In any system that uses structured knowledgegraph (KG) data as its underlying knowledge representati...
Providing pretrained language models with simple task descriptions in natural language enables them ...
Controlling neural network-based models for natural language generation (NLG) to realize desirable a...
Existing solutions to zero-shot text classification either conduct prompting with pre-trained langu...
In data-to-text (D2T) generation, training on in-domain data leads to overfitting to the data repres...
We study the problem of generating a training-free task-dependent visual classifier from text descri...
Our research focuses on solving the zero-shot text classification problem in NLP, with a particular ...
We present a deep generative model for Zero-Shot Learning (ZSL). Unlike most existing methods for th...
Nowadays, owing to the superior capacity of the large pre-trained language models (PLM), the PLM-bas...
Pretrained language models (PLMs) have demonstrated remarkable performance in various natural langua...
In recent years, the community of natural language processing (NLP) has seen amazing progress in the...
International audienceLarge language models have recently been shown to attain reasonable zero-shot ...
Traditional text classification approaches often require a good amount of labeled data, which is dif...
One of the most impressive results of recent NLP history is the ability of pre-trained language mode...
We propose a novel framework for zero-shot learning of topic-dependent language models, which enable...
In any system that uses structured knowledgegraph (KG) data as its underlying knowledge representati...
Providing pretrained language models with simple task descriptions in natural language enables them ...
Controlling neural network-based models for natural language generation (NLG) to realize desirable a...
Existing solutions to zero-shot text classification either conduct prompting with pre-trained langu...
In data-to-text (D2T) generation, training on in-domain data leads to overfitting to the data repres...
We study the problem of generating a training-free task-dependent visual classifier from text descri...
Our research focuses on solving the zero-shot text classification problem in NLP, with a particular ...
We present a deep generative model for Zero-Shot Learning (ZSL). Unlike most existing methods for th...