Prompting has shown impressive success in enabling large pretrained language models (LMs) to perform diverse NLP tasks, especially when only few downstream data are available. Automatically finding the optimal prompt for each task, however, is challenging. Most existing work resorts to tuning soft prompt (e.g., embeddings) which falls short of interpretability, reusability across LMs, and applicability when gradients are not accessible. Discrete prompt, on the other hand, is difficult to optimize, and is often created by "enumeration (e.g., paraphrasing)-then-selection" heuristics that do not explore the prompt space systematically. This paper proposes RLPrompt, an efficient discrete prompt optimization approach with reinforcement learning ...
Pretrained language models (PLMs) have demonstrated remarkable performance in various natural langua...
Pretrained language models (PLMs) have made remarkable progress in text generation tasks via fine-tu...
Why can pre-trained language models (PLMs) learn universal representations and effectively adapt to ...
Reinforcement learning (RL) has emerged as a powerful paradigm for fine-tuning Large Language Models...
Recent works have shown promising results of prompt tuning in stimulating pre-trained language model...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
In recent years, there has been significant progress in developing pre-trained language models for N...
We tackle the problem of aligning pre-trained large language models (LMs) with human preferences. If...
Prompt tuning learns soft prompts to condition frozen Pre-trained Language Models (PLMs) for perform...
Large-scale pre-trained language models have contributed significantly to natural language processin...
Reinforcement learning (RL) has been widely used to aid training in language generation. This is ach...
Prompts have been shown to be an effective method to adapt a frozen Pretrained Language Model (PLM) ...
Usage of reinforcement learning (RL) in natural language processing (NLP) tasks has gained momentum ...
Maximum likelihood estimation (MLE) is the predominant algorithm for training text generation models...
Text style transfer is an important task in controllable language generation. Supervised approaches ...
Pretrained language models (PLMs) have demonstrated remarkable performance in various natural langua...
Pretrained language models (PLMs) have made remarkable progress in text generation tasks via fine-tu...
Why can pre-trained language models (PLMs) learn universal representations and effectively adapt to ...
Reinforcement learning (RL) has emerged as a powerful paradigm for fine-tuning Large Language Models...
Recent works have shown promising results of prompt tuning in stimulating pre-trained language model...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
In recent years, there has been significant progress in developing pre-trained language models for N...
We tackle the problem of aligning pre-trained large language models (LMs) with human preferences. If...
Prompt tuning learns soft prompts to condition frozen Pre-trained Language Models (PLMs) for perform...
Large-scale pre-trained language models have contributed significantly to natural language processin...
Reinforcement learning (RL) has been widely used to aid training in language generation. This is ach...
Prompts have been shown to be an effective method to adapt a frozen Pretrained Language Model (PLM) ...
Usage of reinforcement learning (RL) in natural language processing (NLP) tasks has gained momentum ...
Maximum likelihood estimation (MLE) is the predominant algorithm for training text generation models...
Text style transfer is an important task in controllable language generation. Supervised approaches ...
Pretrained language models (PLMs) have demonstrated remarkable performance in various natural langua...
Pretrained language models (PLMs) have made remarkable progress in text generation tasks via fine-tu...
Why can pre-trained language models (PLMs) learn universal representations and effectively adapt to ...