Large-scale retrieval is to recall relevant documents from a huge collection given a query. It relies on representation learning to embed documents and queries into a common semantic encoding space. According to the encoding space, recent retrieval methods based on pre-trained language models (PLM) can be coarsely categorized into either dense-vector or lexicon-based paradigms. These two paradigms unveil the PLMs' representation capability in different granularities, i.e., global sequence-level compression and local word-level contexts, respectively. Inspired by their complementary global-local contextualization and distinct representing views, we propose a new learning framework, UnifieR, which unifies dense-vector and lexicon-based retrie...
Pseudo Relevance Feedback (PRF) is known to improve the effectiveness of bag-of-words retrievers. At...
The aims of this paper are twofold. Our first aim is to compare results of the earlier Terabyte trac...
Existing Text-to-SQL generators require the entire schema to be encoded with the user text. This is ...
In this work, we propose a simple method that applies a large language model (LLM) to large-scale re...
The advent of contextualised language models has brought gains in search effectiveness, not just whe...
Dual-encoder-based dense retrieval models have become the standard in IR. They employ large Transfor...
In this paper, we describe a classifier based retrieval scheme for efficiently and accurately retrie...
Here the models described in the publication "PARM: A Paragraph Aggregation Retrieval Model for Dens...
Deeply learned representations have achieved superior image retrieval performance in a retrieve-then...
Recently, retrieval models based on dense representations are dominant in passage retrieval tasks, d...
On a wide range of natural language processing and information retrieval tasks, transformer-based mo...
Information retrieval is an important task that requires specific attention in the biomedical domain...
Despite their recent popularity and well-known advantages, dense retrievers still lag behind sparse ...
Deep pre-trained language models (e,g. BERT) are effective at large-scale text retrieval task. Exist...
Although information access systems have long supported people in accomplishing a wide range of task...
Pseudo Relevance Feedback (PRF) is known to improve the effectiveness of bag-of-words retrievers. At...
The aims of this paper are twofold. Our first aim is to compare results of the earlier Terabyte trac...
Existing Text-to-SQL generators require the entire schema to be encoded with the user text. This is ...
In this work, we propose a simple method that applies a large language model (LLM) to large-scale re...
The advent of contextualised language models has brought gains in search effectiveness, not just whe...
Dual-encoder-based dense retrieval models have become the standard in IR. They employ large Transfor...
In this paper, we describe a classifier based retrieval scheme for efficiently and accurately retrie...
Here the models described in the publication "PARM: A Paragraph Aggregation Retrieval Model for Dens...
Deeply learned representations have achieved superior image retrieval performance in a retrieve-then...
Recently, retrieval models based on dense representations are dominant in passage retrieval tasks, d...
On a wide range of natural language processing and information retrieval tasks, transformer-based mo...
Information retrieval is an important task that requires specific attention in the biomedical domain...
Despite their recent popularity and well-known advantages, dense retrievers still lag behind sparse ...
Deep pre-trained language models (e,g. BERT) are effective at large-scale text retrieval task. Exist...
Although information access systems have long supported people in accomplishing a wide range of task...
Pseudo Relevance Feedback (PRF) is known to improve the effectiveness of bag-of-words retrievers. At...
The aims of this paper are twofold. Our first aim is to compare results of the earlier Terabyte trac...
Existing Text-to-SQL generators require the entire schema to be encoded with the user text. This is ...