Pre-training on larger datasets with ever increasing model size is now a proven recipe for increased performance across almost all NLP tasks. A notable exception is information retrieval, where additional pre-training has so far failed to produce convincing results. We show that, with the right pre-training setup, this barrier can be overcome. We demonstrate this by pre-training large bi-encoder models on 1) a recently released set of 65 million synthetically generated questions, and 2) 200 million post-comment pairs from a preexisting dataset of Reddit conversations. We evaluate on a set of information retrieval and dialogue retrieval benchmarks, showing substantial improvements over supervised baselines
Pretrained language models have become the standard approach for many NLP tasks due to strong perfor...
Existing pre-trained models are generally geared towards a particular class of problems. To date, th...
Dense retrieval (DR) converts queries and documents into dense embeddings and measures the similarit...
Dense retrievers for open-domain question answering (ODQA) have been shown to achieve impressive per...
Recently, methods have been developed to improve the performance of dense passage retrieval by using...
Thesis (Ph.D.)--University of Washington, 2022A robust language processing machine should be able to...
The pre-training and fine-tuning paradigm has contributed to a number of breakthroughs in Natural La...
This paper studies multi-task training of retrieval-augmented generation models for knowledge-intens...
Recent progress in pretraining language models on large textual corpora led to a surge of improvemen...
Recent advances in large-scale pre-training provide large models with the potential to learn knowled...
Dense retrieval models have predominantly been studied for English, where models have shown great su...
As the demand for sophisticated Natural Language Processing (NLP) models continues to grow, so does ...
Recently, the development of pre-trained language models has brought natural language processing (NL...
Finetuning Pretrained Language Models (PLM) for IR has been de facto the standard practice since the...
Bi-encoders and cross-encoders are widely used in many state-of-the-art retrieval pipelines. In this...
Pretrained language models have become the standard approach for many NLP tasks due to strong perfor...
Existing pre-trained models are generally geared towards a particular class of problems. To date, th...
Dense retrieval (DR) converts queries and documents into dense embeddings and measures the similarit...
Dense retrievers for open-domain question answering (ODQA) have been shown to achieve impressive per...
Recently, methods have been developed to improve the performance of dense passage retrieval by using...
Thesis (Ph.D.)--University of Washington, 2022A robust language processing machine should be able to...
The pre-training and fine-tuning paradigm has contributed to a number of breakthroughs in Natural La...
This paper studies multi-task training of retrieval-augmented generation models for knowledge-intens...
Recent progress in pretraining language models on large textual corpora led to a surge of improvemen...
Recent advances in large-scale pre-training provide large models with the potential to learn knowled...
Dense retrieval models have predominantly been studied for English, where models have shown great su...
As the demand for sophisticated Natural Language Processing (NLP) models continues to grow, so does ...
Recently, the development of pre-trained language models has brought natural language processing (NL...
Finetuning Pretrained Language Models (PLM) for IR has been de facto the standard practice since the...
Bi-encoders and cross-encoders are widely used in many state-of-the-art retrieval pipelines. In this...
Pretrained language models have become the standard approach for many NLP tasks due to strong perfor...
Existing pre-trained models are generally geared towards a particular class of problems. To date, th...
Dense retrieval (DR) converts queries and documents into dense embeddings and measures the similarit...