Current efficient fine-tuning methods (e.g., adapters, prefix-tuning, etc.) have optimized conditional text generation via training a small set of extra parameters of the neural language model, while freezing the rest for efficiency. While showing strong performance on some generation tasks, they don't generalize across all generation tasks. We show that soft-prompt based conditional text generation can be improved with simple and efficient methods that simulate modeling the discourse structure of human written text. We investigate two design choices: First, we apply \textit{hierarchical blocking} on the prefix parameters to simulate a higher-level discourse structure of human written text. Second, we apply \textit{attention sparsity} on th...
Neural language models often fail to generate diverse and informative texts, limiting their applicab...
Few-shot abstractive summarization has become a challenging task in natural language generation. To ...
Following some recent propositions to handle natural language generation in spoken dialogue systems ...
Prefix-tuning is a powerful lightweight technique for adapting a large pre-trained language model to...
The current sequence-to-sequence with attention models, despite being successful, are inherently lim...
Speech representations learned from Self-supervised learning (SSL) models can benefit various speech...
Prompt tuning attempts to update few task-specific parameters in pre-trained models. It has achieved...
Large language models (LM) based on Transformers allow to generate plausible long texts. In this pap...
We explore the idea of compressing the prompts used to condition language models, and show that comp...
Language model fine-tuning is essential for modern natural language processing, but is computational...
Deep neural language models like GPT-2 is undoubtedly strong at text generation, but often requires ...
Large pre-trained language models have repeatedly shown their ability to produce fluent text. Yet ev...
Providing pretrained language models with simple task descriptions in natural language enables them ...
Pretrained Transformer-based language models (LMs) display remarkable natural language generation ca...
Controllable text generation systems often leverage control codes to direct various properties of th...
Neural language models often fail to generate diverse and informative texts, limiting their applicab...
Few-shot abstractive summarization has become a challenging task in natural language generation. To ...
Following some recent propositions to handle natural language generation in spoken dialogue systems ...
Prefix-tuning is a powerful lightweight technique for adapting a large pre-trained language model to...
The current sequence-to-sequence with attention models, despite being successful, are inherently lim...
Speech representations learned from Self-supervised learning (SSL) models can benefit various speech...
Prompt tuning attempts to update few task-specific parameters in pre-trained models. It has achieved...
Large language models (LM) based on Transformers allow to generate plausible long texts. In this pap...
We explore the idea of compressing the prompts used to condition language models, and show that comp...
Language model fine-tuning is essential for modern natural language processing, but is computational...
Deep neural language models like GPT-2 is undoubtedly strong at text generation, but often requires ...
Large pre-trained language models have repeatedly shown their ability to produce fluent text. Yet ev...
Providing pretrained language models with simple task descriptions in natural language enables them ...
Pretrained Transformer-based language models (LMs) display remarkable natural language generation ca...
Controllable text generation systems often leverage control codes to direct various properties of th...
Neural language models often fail to generate diverse and informative texts, limiting their applicab...
Few-shot abstractive summarization has become a challenging task in natural language generation. To ...
Following some recent propositions to handle natural language generation in spoken dialogue systems ...