The dominant approaches for controlling language models achieve prominence in controlling high-level attributes (e.g. topic and sentiment). However, these methods often require condition-specific data or are computationally expensive. We propose a new simple guided decoding method, Gamma Sampling, which does not require any training data to achieve fine-grained controllable text generation while maintaining a fast generation speed. Gamma Sampling introduces attribute-related information (provided by humans or language models themselves) into the sampling process to guide language models to generate texts with desired attributes. Since no training is involved, Gamma Sampling can be easily applied to any language model for controllable text g...
Scaling language models with more data, compute and parameters has driven significant progress in na...
In recent years, the field of language modelling has witnessed exciting developments. Especially, th...
Neural language models are increasingly deployed into APIs and websites that allow a user to pass in...
Controllable text generation (CTG) aims to generate text with desired attributes, and decoding-time-...
Language model fine-tuning is essential for modern natural language processing, but is computational...
Pretrained language models have demonstrated extraordinary capabilities in language generation. Howe...
Controllable text generation systems often leverage control codes to direct various properties of th...
Existing reference-free metrics have obvious limitations for evaluating controlled text generation m...
High-quality instruction-tuning data is critical to improving LLM capabilities. Existing data collec...
Despite considerable advances in neural language modeling, it remains an open question what the best...
We explore the idea of compressing the prompts used to condition language models, and show that comp...
Pretrained Transformer-based language models (LMs) display remarkable natural language generation ca...
Prefix-tuning is a powerful lightweight technique for adapting a large pre-trained language model to...
Long samples of text from neural language models can be of poor quality. Truncation sampling algorit...
Today's probabilistic language generators fall short when it comes to producing coherent and fluent ...
Scaling language models with more data, compute and parameters has driven significant progress in na...
In recent years, the field of language modelling has witnessed exciting developments. Especially, th...
Neural language models are increasingly deployed into APIs and websites that allow a user to pass in...
Controllable text generation (CTG) aims to generate text with desired attributes, and decoding-time-...
Language model fine-tuning is essential for modern natural language processing, but is computational...
Pretrained language models have demonstrated extraordinary capabilities in language generation. Howe...
Controllable text generation systems often leverage control codes to direct various properties of th...
Existing reference-free metrics have obvious limitations for evaluating controlled text generation m...
High-quality instruction-tuning data is critical to improving LLM capabilities. Existing data collec...
Despite considerable advances in neural language modeling, it remains an open question what the best...
We explore the idea of compressing the prompts used to condition language models, and show that comp...
Pretrained Transformer-based language models (LMs) display remarkable natural language generation ca...
Prefix-tuning is a powerful lightweight technique for adapting a large pre-trained language model to...
Long samples of text from neural language models can be of poor quality. Truncation sampling algorit...
Today's probabilistic language generators fall short when it comes to producing coherent and fluent ...
Scaling language models with more data, compute and parameters has driven significant progress in na...
In recent years, the field of language modelling has witnessed exciting developments. Especially, th...
Neural language models are increasingly deployed into APIs and websites that allow a user to pass in...