Due to high annotation costs making the best use of existing human-created training data is an important research direction. We, therefore, carry out a systematic evaluation of transferability of BERT-based neural ranking models across five English datasets. Previous studies focused primarily on zero-shot and few-shot transfer from a large dataset to a dataset with a small number of queries. In contrast, each of our collections has a substantial number of queries, which enables a full-shot evaluation mode and improves reliability of our results. Furthermore, since source datasets licences often prohibit commercial use, we compare transfer learning to training on pseudo-labels generated by a BM25 scorer. We find that training on pseudo-label...
The limited availability of ground truth relevance labels has been a major impediment to the applica...
Technology-assisted review (TAR) refers to iterative active learning workflows for document review i...
The emergence of BERT in 2018 has brought a huge boon to retrieval effectiveness in many tasks acros...
[Background]: The advent of bidirectional encoder representation from trans- formers (BERT) language...
We propose utilizing n-best reranking to enhance the Sequence-Level Knowledge Distillation (Kim and ...
Data annotation is the process of labeling text, images, or other types of content for machine learn...
In recent years, large pre-trained transformers have led to substantial gains in performance over tr...
In a great deal of theoretical and applied cognitive and neurophysiological research, it is essentia...
A lack of reliable relevance labels for training ranking functions is a significant problem for many...
The lack of relevance labels is increasingly challenging and presents a bottleneck in the training o...
One challenge with neural ranking is the need for a large amount of manually-labeled relevance judgm...
Learning-to-Rank (LtR) techniques leverage machine learning algorithms and large amounts of training...
Over the past decade, the field of machine learning has experienced remarkable advancements. While i...
Pre-training language models (LMs) on large-scale unlabeled text data makes the model much easier to...
Learning to rank techniques provide mechanisms for combining document feature values into learned mo...
The limited availability of ground truth relevance labels has been a major impediment to the applica...
Technology-assisted review (TAR) refers to iterative active learning workflows for document review i...
The emergence of BERT in 2018 has brought a huge boon to retrieval effectiveness in many tasks acros...
[Background]: The advent of bidirectional encoder representation from trans- formers (BERT) language...
We propose utilizing n-best reranking to enhance the Sequence-Level Knowledge Distillation (Kim and ...
Data annotation is the process of labeling text, images, or other types of content for machine learn...
In recent years, large pre-trained transformers have led to substantial gains in performance over tr...
In a great deal of theoretical and applied cognitive and neurophysiological research, it is essentia...
A lack of reliable relevance labels for training ranking functions is a significant problem for many...
The lack of relevance labels is increasingly challenging and presents a bottleneck in the training o...
One challenge with neural ranking is the need for a large amount of manually-labeled relevance judgm...
Learning-to-Rank (LtR) techniques leverage machine learning algorithms and large amounts of training...
Over the past decade, the field of machine learning has experienced remarkable advancements. While i...
Pre-training language models (LMs) on large-scale unlabeled text data makes the model much easier to...
Learning to rank techniques provide mechanisms for combining document feature values into learned mo...
The limited availability of ground truth relevance labels has been a major impediment to the applica...
Technology-assisted review (TAR) refers to iterative active learning workflows for document review i...
The emergence of BERT in 2018 has brought a huge boon to retrieval effectiveness in many tasks acros...