Retrieval with extremely long queries and documents is a well-known and challenging task in information retrieval and is commonly known as Query-by-Document (QBD) retrieval. Specifically designed Transformer models that can handle long input sequences have not shown high effectiveness in QBD tasks in previous work. We propose a Re-Ranker based on the novel Proportional Relevance Score (RPRS) to compute the relevance score between a query and the top-k candidate documents. Our extensive evaluation shows RPRS obtains significantly better results than the state-of-the-art models on five different datasets. Furthermore, RPRS is highly efficient since all documents can be pre-processed, embedded, and indexed before query time which gives our re-...
Recent work has shown that small distilled language models are strong competitors to models that are...
Recent research has shown that transformer networks can be used as differentiable search indexes by ...
Query rewriting plays a vital role in enhancing conversational search by transforming context-depend...
Query-by-document (QBD) retrieval is an Information Retrieval task in which a seed document acts as ...
Dual-encoder-based dense retrieval models have become the standard in IR. They employ large Transfor...
On a wide range of natural language processing and information retrieval tasks, transformer-based mo...
Neural ranking methods based on large transformer models have recently gained significant attention ...
Deep pre-trained language models (e,g. BERT) are effective at large-scale text retrieval task. Exist...
Many information retrieval tasks require large labeled datasets for fine-tuning. However, such datas...
International audienceOn a wide range of natural language processing and information retrieval tasks...
International audienceTransformer-based models, and especially pre-trained language models like BERT...
Dense retrieval, which describes the use of contextualised language models such as BERT to identify ...
In this work, we propose a simple method that applies a large language model (LLM) to large-scale re...
A long query provides more useful hints for searching relevant documents, but it is likely to introd...
Fang, HuiThe development of information retrieval (IR) (the search engine) is one of the revolutiona...
Recent work has shown that small distilled language models are strong competitors to models that are...
Recent research has shown that transformer networks can be used as differentiable search indexes by ...
Query rewriting plays a vital role in enhancing conversational search by transforming context-depend...
Query-by-document (QBD) retrieval is an Information Retrieval task in which a seed document acts as ...
Dual-encoder-based dense retrieval models have become the standard in IR. They employ large Transfor...
On a wide range of natural language processing and information retrieval tasks, transformer-based mo...
Neural ranking methods based on large transformer models have recently gained significant attention ...
Deep pre-trained language models (e,g. BERT) are effective at large-scale text retrieval task. Exist...
Many information retrieval tasks require large labeled datasets for fine-tuning. However, such datas...
International audienceOn a wide range of natural language processing and information retrieval tasks...
International audienceTransformer-based models, and especially pre-trained language models like BERT...
Dense retrieval, which describes the use of contextualised language models such as BERT to identify ...
In this work, we propose a simple method that applies a large language model (LLM) to large-scale re...
A long query provides more useful hints for searching relevant documents, but it is likely to introd...
Fang, HuiThe development of information retrieval (IR) (the search engine) is one of the revolutiona...
Recent work has shown that small distilled language models are strong competitors to models that are...
Recent research has shown that transformer networks can be used as differentiable search indexes by ...
Query rewriting plays a vital role in enhancing conversational search by transforming context-depend...