International audienceLarge language models have recently been shown to attain reasonable zero-shot generalization on a diverse set of tasks (Brown et al., 2020). It has been hypothesized that this is a consequence of implicit multitask learning in language models’ pretraining (Radford et al., 2019). Can zero-shot generalization instead be directly induced by explicit multitask learning? To test this question at scale, we develop a system for easily mapping any natural language tasks into a human-readable prompted form. We convert a large set of supervised datasets, each with multiple prompts with diverse wording. These prompted datasets allow for benchmarking the ability of a model to perform completely held-out tasks. We fine-tune a pre-t...
Large language models readily adapt to novel settings, even without task-specific training data. Can...
Recent research has shown promise in multilingual modeling, demonstrating how a single model is capa...
Providing pretrained language models with simple task descriptions in natural language enables them ...
International audienceLarge language models have recently been shown to attain reasonable zero-shot ...
One of the most impressive results of recent NLP history is the ability of pre-trained language mode...
We propose a multitask pretraining approach ZeroPrompt for zero-shot generalization, focusing on tas...
Large-scale pre-trained language models have contributed significantly to natural language processin...
Large-scale generative language models such as GPT-3 are competitive few-shot learners. While these ...
Instruction tuning is an emergent paradigm in NLP wherein natural language instructions are leverage...
Pretrained language models (PLMs) have demonstrated remarkable performance in various natural langua...
Deep learning has recently driven remarkable progress in several applications, including image class...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
There is a growing interest in dataset generation recently due to the superior generative capacity o...
Most combinations of NLP tasks and language varieties lack in-domain examples for supervised trainin...
Pretraining deep neural networks to perform language modeling - that is, to reconstruct missing word...
Large language models readily adapt to novel settings, even without task-specific training data. Can...
Recent research has shown promise in multilingual modeling, demonstrating how a single model is capa...
Providing pretrained language models with simple task descriptions in natural language enables them ...
International audienceLarge language models have recently been shown to attain reasonable zero-shot ...
One of the most impressive results of recent NLP history is the ability of pre-trained language mode...
We propose a multitask pretraining approach ZeroPrompt for zero-shot generalization, focusing on tas...
Large-scale pre-trained language models have contributed significantly to natural language processin...
Large-scale generative language models such as GPT-3 are competitive few-shot learners. While these ...
Instruction tuning is an emergent paradigm in NLP wherein natural language instructions are leverage...
Pretrained language models (PLMs) have demonstrated remarkable performance in various natural langua...
Deep learning has recently driven remarkable progress in several applications, including image class...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
There is a growing interest in dataset generation recently due to the superior generative capacity o...
Most combinations of NLP tasks and language varieties lack in-domain examples for supervised trainin...
Pretraining deep neural networks to perform language modeling - that is, to reconstruct missing word...
Large language models readily adapt to novel settings, even without task-specific training data. Can...
Recent research has shown promise in multilingual modeling, demonstrating how a single model is capa...
Providing pretrained language models with simple task descriptions in natural language enables them ...