Zero-shot cross-domain slot filling aims to transfer knowledge from the labeled source domain to the unlabeled target domain. Existing models either encode slot descriptions and examples or design handcrafted question templates using heuristic rules, suffering from poor generalization capability or robustness. In this paper, we propose a generative zero-shot prompt learning framework for cross-domain slot filling, both improving generalization and robustness than previous work. Besides, we introduce a novel inverse prompting strategy to distinguish different slot types to avoid the multiple prediction problem, and an efficient prompt-tuning strategy to boost higher performance by only training fewer prompt parameters. Experiments and analys...
There is a growing interest in dataset generation recently due to the superior generative capacity o...
Large pre-trained vision-language models like CLIP have shown great potential in learning representa...
Large-scale pre-trained language models have contributed significantly to natural language processin...
This paper addresses zero-shot slot filling, which tries to build a system that can generalize to un...
International audienceLarge language models have recently been shown to attain reasonable zero-shot ...
We propose a multitask pretraining approach ZeroPrompt for zero-shot generalization, focusing on tas...
Pretrained language models have shown success in various areas of natural language processing, inclu...
One of the most impressive results of recent NLP history is the ability of pre-trained language mode...
Conversational agents such as Alexa and Google Assistant constantly need to increase their language ...
Few-shot slot tagging is an important task in dialogue systems and attracts much attention of resear...
Slot filling is a core operation for utterance understanding in task-oriented dialogue systems. Slot...
NLP has yielded results that were unimaginable only a few years ago on a wide range of real-world ta...
We propose a simple and effective re-ranking method for improving passage retrieval in open question...
Zero-shot slot filling has received considerable attention to cope with the problem of limited avail...
Manually annotating fine-grained slot-value labels for task-oriented dialogue (ToD) systems is an ex...
There is a growing interest in dataset generation recently due to the superior generative capacity o...
Large pre-trained vision-language models like CLIP have shown great potential in learning representa...
Large-scale pre-trained language models have contributed significantly to natural language processin...
This paper addresses zero-shot slot filling, which tries to build a system that can generalize to un...
International audienceLarge language models have recently been shown to attain reasonable zero-shot ...
We propose a multitask pretraining approach ZeroPrompt for zero-shot generalization, focusing on tas...
Pretrained language models have shown success in various areas of natural language processing, inclu...
One of the most impressive results of recent NLP history is the ability of pre-trained language mode...
Conversational agents such as Alexa and Google Assistant constantly need to increase their language ...
Few-shot slot tagging is an important task in dialogue systems and attracts much attention of resear...
Slot filling is a core operation for utterance understanding in task-oriented dialogue systems. Slot...
NLP has yielded results that were unimaginable only a few years ago on a wide range of real-world ta...
We propose a simple and effective re-ranking method for improving passage retrieval in open question...
Zero-shot slot filling has received considerable attention to cope with the problem of limited avail...
Manually annotating fine-grained slot-value labels for task-oriented dialogue (ToD) systems is an ex...
There is a growing interest in dataset generation recently due to the superior generative capacity o...
Large pre-trained vision-language models like CLIP have shown great potential in learning representa...
Large-scale pre-trained language models have contributed significantly to natural language processin...