[Background]: The advent of bidirectional encoder representation from trans- formers (BERT) language models (Devlin et al., 2018) and MS Marco, a large scale human-annotated dataset for machine reading comprehension (Bajaj et al., 2016) that made publicly available, led the field of information retrieval (IR) to experience a revolution (Lin et al., 2020). The retrieval model based on BERT of Nogueira and Cho (2019), by the time they published their paper, became the top entry in the MS Marco passage-reranking leaderboard, surpassing the previous state of the art by 27% in MRR@10. However, training such neural IR models for different domains than MS Marco is still hard because neural approaches often require a vast amount of training data to...
This thesis work is in the fields of textual information retrieval (IR) and deep learning using neur...
Neural networks with deep architectures have demonstrated significant performance improvements in co...
Transformer-based pre-trained language models like BERT and its variants have recently achieved prom...
[Background]: The advent of bidirectional encoder representation from trans- formers (BERT) language...
Due to high annotation costs making the best use of existing human-created training data is an impor...
Recent developments of machine learning models, and in particular deep neural networks, have yielded...
Recent developments in neural information retrieval models have been promising, but a problem remain...
In recent years, large pre-trained transformers have led to substantial gains in performance over tr...
Information Retriever (IR) aims to find the relevant documents (e.g. snippets, passages, and article...
International audienceThe Information Retrieval (IR) community has witnessed a flourishing developme...
Neural ranking methods based on large transformer models have recently gained significant attention ...
Heavily pre-trained transformers for language modeling, such as BERT, have shown to be remarkably ef...
Deep neural language models such as BERT have enabled substantial recent advances in many natural la...
When pre-trained on large unsupervised textual corpora, language models are able to store and retri...
Technology-assisted review (TAR) refers to iterative active learning workflows for document review i...
This thesis work is in the fields of textual information retrieval (IR) and deep learning using neur...
Neural networks with deep architectures have demonstrated significant performance improvements in co...
Transformer-based pre-trained language models like BERT and its variants have recently achieved prom...
[Background]: The advent of bidirectional encoder representation from trans- formers (BERT) language...
Due to high annotation costs making the best use of existing human-created training data is an impor...
Recent developments of machine learning models, and in particular deep neural networks, have yielded...
Recent developments in neural information retrieval models have been promising, but a problem remain...
In recent years, large pre-trained transformers have led to substantial gains in performance over tr...
Information Retriever (IR) aims to find the relevant documents (e.g. snippets, passages, and article...
International audienceThe Information Retrieval (IR) community has witnessed a flourishing developme...
Neural ranking methods based on large transformer models have recently gained significant attention ...
Heavily pre-trained transformers for language modeling, such as BERT, have shown to be remarkably ef...
Deep neural language models such as BERT have enabled substantial recent advances in many natural la...
When pre-trained on large unsupervised textual corpora, language models are able to store and retri...
Technology-assisted review (TAR) refers to iterative active learning workflows for document review i...
This thesis work is in the fields of textual information retrieval (IR) and deep learning using neur...
Neural networks with deep architectures have demonstrated significant performance improvements in co...
Transformer-based pre-trained language models like BERT and its variants have recently achieved prom...