Prompt tuning attempts to update few task-specific parameters in pre-trained models. It has achieved comparable performance to fine-tuning of the full parameter set on both language understanding and generation tasks. In this work, we study the problem of prompt tuning for neural text retrievers. We introduce parameter-efficient prompt tuning for text retrieval across in-domain, cross-domain, and cross-topic settings. Through an extensive analysis, we show that the strategy can mitigate the two issues -- parameter-inefficiency and weak generalizability -- faced by fine-tuning based retrieval methods. Notably, it can significantly improve the out-of-domain zero-shot generalization of the retrieval models. By updating only 0.1% of the model p...
Automatic Speech Recognition (ASR) systems have found their use in numerous industrial applications ...
Fine-grained object retrieval aims to learn discriminative representation to retrieve visually simil...
Enhancing the zero-shot performance of instruction-following models requires heavy computation, eith...
While GPTs with traditional fine-tuning fail to achieve strong results on natural language understan...
Prompt tuning has become a new paradigm for model tuning and it has demonstrated success in natural ...
Current efficient fine-tuning methods (e.g., adapters, prefix-tuning, etc.) have optimized condition...
Pre-trained multilingual language models show significant performance gains for zero-shot cross-ling...
Dense retrieval (DR) converts queries and documents into dense embeddings and measures the similarit...
Prefix-tuning is a powerful lightweight technique for adapting a large pre-trained language model to...
The advent of hyper-scale and general-purpose pre-trained models is shifting the paradigm of buildin...
Language model fine-tuning is essential for modern natural language processing, but is computational...
Prompt tuning learns soft prompts to condition frozen Pre-trained Language Models (PLMs) for perform...
Recent studies show that prompt tuning can better leverage the power of large language models than f...
Speech representations learned from Self-supervised learning (SSL) models can benefit various speech...
In recent years, prompt tuning has sparked a research surge in adapting pre-trained models. Unlike t...
Automatic Speech Recognition (ASR) systems have found their use in numerous industrial applications ...
Fine-grained object retrieval aims to learn discriminative representation to retrieve visually simil...
Enhancing the zero-shot performance of instruction-following models requires heavy computation, eith...
While GPTs with traditional fine-tuning fail to achieve strong results on natural language understan...
Prompt tuning has become a new paradigm for model tuning and it has demonstrated success in natural ...
Current efficient fine-tuning methods (e.g., adapters, prefix-tuning, etc.) have optimized condition...
Pre-trained multilingual language models show significant performance gains for zero-shot cross-ling...
Dense retrieval (DR) converts queries and documents into dense embeddings and measures the similarit...
Prefix-tuning is a powerful lightweight technique for adapting a large pre-trained language model to...
The advent of hyper-scale and general-purpose pre-trained models is shifting the paradigm of buildin...
Language model fine-tuning is essential for modern natural language processing, but is computational...
Prompt tuning learns soft prompts to condition frozen Pre-trained Language Models (PLMs) for perform...
Recent studies show that prompt tuning can better leverage the power of large language models than f...
Speech representations learned from Self-supervised learning (SSL) models can benefit various speech...
In recent years, prompt tuning has sparked a research surge in adapting pre-trained models. Unlike t...
Automatic Speech Recognition (ASR) systems have found their use in numerous industrial applications ...
Fine-grained object retrieval aims to learn discriminative representation to retrieve visually simil...
Enhancing the zero-shot performance of instruction-following models requires heavy computation, eith...