This work studies a challenging yet more realistic setting for zero-shot cross-task generalization: demonstration-free learning from textual instructions, presuming the existence of a paragraph-style task definition while no demonstrations exist. To better learn the task supervision from the definition, we propose two strategies: first, to automatically find out the critical sentences in the definition; second, a ranking objective to force the model to generate the gold outputs with higher probabilities when those critical parts are highlighted in the definition. The joint efforts of the two strategies yield state-of-the-art performance on the challenging benchmark. Our code will be released in the final version of the paper.Comment: Prepri...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
We propose a multitask pretraining approach ZeroPrompt for zero-shot generalization, focusing on tas...
Can we teach a robot to recognize and make predictions for activities that it has never seen before?...
One of the most impressive results of recent NLP history is the ability of pre-trained language mode...
International audienceLarge language models have recently been shown to attain reasonable zero-shot ...
High-quality instruction-tuning data is critical to improving LLM capabilities. Existing data collec...
Instruction tuning is an emergent paradigm in NLP wherein natural language instructions are leverage...
Large language models are able to perform a task by conditioning on a few input-output demonstration...
Providing pretrained language models with simple task descriptions in natural language enables them ...
This paper presents AutoHint, a novel framework for automatic prompt engineering and optimization fo...
Recent work has shown that fine-tuning large pre-trained language models on a collection of tasks de...
Pretrained language models (PLMs) have demonstrated remarkable performance in various natural langua...
Large language models (LMs) are able to in-context learn -- perform a new task via inference alone b...
In this work, we evaluate 10 open-source instructed LLMs on four representative code comprehension a...
This paper studies the use of language models as a source of synthetic unlabeled text for NLP. We fo...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
We propose a multitask pretraining approach ZeroPrompt for zero-shot generalization, focusing on tas...
Can we teach a robot to recognize and make predictions for activities that it has never seen before?...
One of the most impressive results of recent NLP history is the ability of pre-trained language mode...
International audienceLarge language models have recently been shown to attain reasonable zero-shot ...
High-quality instruction-tuning data is critical to improving LLM capabilities. Existing data collec...
Instruction tuning is an emergent paradigm in NLP wherein natural language instructions are leverage...
Large language models are able to perform a task by conditioning on a few input-output demonstration...
Providing pretrained language models with simple task descriptions in natural language enables them ...
This paper presents AutoHint, a novel framework for automatic prompt engineering and optimization fo...
Recent work has shown that fine-tuning large pre-trained language models on a collection of tasks de...
Pretrained language models (PLMs) have demonstrated remarkable performance in various natural langua...
Large language models (LMs) are able to in-context learn -- perform a new task via inference alone b...
In this work, we evaluate 10 open-source instructed LLMs on four representative code comprehension a...
This paper studies the use of language models as a source of synthetic unlabeled text for NLP. We fo...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
We propose a multitask pretraining approach ZeroPrompt for zero-shot generalization, focusing on tas...
Can we teach a robot to recognize and make predictions for activities that it has never seen before?...