Black-Box Tuning (BBT) is a derivative-free approach to optimize continuous prompt tokens prepended to the input of language models. Although BBT has achieved comparable performance to full model tuning on simple classification tasks under few-shot settings, it requires pre-trained prompt embedding to match model tuning on hard tasks (e.g., entailment tasks), and therefore does not completely get rid of the dependence on gradients. In this paper we present BBTv2, a pure black-box optimization approach that can drive language models to achieve comparable results to gradient-based optimization. In particular, we prepend continuous prompt tokens to every layer of the language model and propose a divide-and-conquer algorithm to alternately opti...
Standard fine-tuning of large pre-trained language models (PLMs) for downstream tasks requires updat...
We present a new method LiST is short for Lite Prompted Self-Training for parameter-efficient fine-t...
Pretrained language models (PLMs) have demonstrated remarkable performance in various natural langua...
Prompt-based learning has been an effective paradigm for large pretrained language models (LLM), ena...
We introduce BitFit, a sparse-finetuning method where only the bias-terms of the model (or a subset ...
The encoding of solutions in black-box optimization is a delicate, handcrafted balance between expre...
Large language models~(LLMs) are instruction followers, but it can be challenging to find the best i...
Through in-context learning (ICL), large-scale language models are effective few-shot learners witho...
Many important problems in science and engineering, such as drug design, involve optimizing an expen...
International audienceExisting studies in black-box optimization for machine learning suffer from lo...
Black box optimization is a field of the global optimization which consists in a family of methods ...
We propose a multitask pretraining approach ZeroPrompt for zero-shot generalization, focusing on tas...
Inspired by the concept of preconditioning, we propose a novel method to increase adaptation speed f...
The diverse world of machine learning applications has given rise to a plethora of algorithms and op...
Nowadays, owing to the superior capacity of the large pre-trained language models (PLM), the PLM-bas...
Standard fine-tuning of large pre-trained language models (PLMs) for downstream tasks requires updat...
We present a new method LiST is short for Lite Prompted Self-Training for parameter-efficient fine-t...
Pretrained language models (PLMs) have demonstrated remarkable performance in various natural langua...
Prompt-based learning has been an effective paradigm for large pretrained language models (LLM), ena...
We introduce BitFit, a sparse-finetuning method where only the bias-terms of the model (or a subset ...
The encoding of solutions in black-box optimization is a delicate, handcrafted balance between expre...
Large language models~(LLMs) are instruction followers, but it can be challenging to find the best i...
Through in-context learning (ICL), large-scale language models are effective few-shot learners witho...
Many important problems in science and engineering, such as drug design, involve optimizing an expen...
International audienceExisting studies in black-box optimization for machine learning suffer from lo...
Black box optimization is a field of the global optimization which consists in a family of methods ...
We propose a multitask pretraining approach ZeroPrompt for zero-shot generalization, focusing on tas...
Inspired by the concept of preconditioning, we propose a novel method to increase adaptation speed f...
The diverse world of machine learning applications has given rise to a plethora of algorithms and op...
Nowadays, owing to the superior capacity of the large pre-trained language models (PLM), the PLM-bas...
Standard fine-tuning of large pre-trained language models (PLMs) for downstream tasks requires updat...
We present a new method LiST is short for Lite Prompted Self-Training for parameter-efficient fine-t...
Pretrained language models (PLMs) have demonstrated remarkable performance in various natural langua...