We propose TandA, an effective technique for fine-tuning pre-trained Transformer models for natural language tasks. Specifically, we first transfer a pre-trained model into a model for a general task by fine-tuning it with a large and high-quality dataset. We then perform a second fine-tuning step to adapt the transferred model to the target domain. We demonstrate the benefits of our approach for answer sentence selection, which is a well-known inference task in Question Answering. We built a large scale dataset to enable the transfer step, exploiting the Natural Questions dataset. Our approach establishes the state of the art on two well-known benchmarks, WikiQA and TREC-QA, achieving the impressive MAP scores of 92% and 94.3%, respectivel...
Natural language processing (NLP) techniques had significantly improved by introducing pre-trained l...
Data annotation is critical for machine learning based natural language processing models. Although ...
The sequence-to-sequence model is a widely used model for dialogue response generators, but it tends...
An important task for designing QA systems is answer sentence selection (AS2): selecting the sentenc...
Transformer models have achieved promising results on natural language processing (NLP) tasks includ...
Inference tasks such as answer sentence selection (AS2) or fact verification are typically solved by...
In Natural Language Processing (NLP), Automatic Question Generation (AQG) is an important task that ...
Answer Sentence Selection is one of the steps typically involved in Question Answering. Question Ans...
El objetivo de este trabajo es proponer nuevas técnicas de Fine-Tuning para mejorar los modelos del ...
Question Answering (QA) systems can significantly reduce manual effort of searching for relevant inf...
We developed a method for producing statistical language models for speech-driven question answering...
Large neural language models are steadily contributing state-of-the-art performance to question answ...
The goal of this article is to develop a multiple-choice questions generation system that has a numb...
Open-domain question answering (QA) is an emerging information-seeking paradigm, which automatically...
Machine Translation models are trained to translate a variety of documents from one language into an...
Natural language processing (NLP) techniques had significantly improved by introducing pre-trained l...
Data annotation is critical for machine learning based natural language processing models. Although ...
The sequence-to-sequence model is a widely used model for dialogue response generators, but it tends...
An important task for designing QA systems is answer sentence selection (AS2): selecting the sentenc...
Transformer models have achieved promising results on natural language processing (NLP) tasks includ...
Inference tasks such as answer sentence selection (AS2) or fact verification are typically solved by...
In Natural Language Processing (NLP), Automatic Question Generation (AQG) is an important task that ...
Answer Sentence Selection is one of the steps typically involved in Question Answering. Question Ans...
El objetivo de este trabajo es proponer nuevas técnicas de Fine-Tuning para mejorar los modelos del ...
Question Answering (QA) systems can significantly reduce manual effort of searching for relevant inf...
We developed a method for producing statistical language models for speech-driven question answering...
Large neural language models are steadily contributing state-of-the-art performance to question answ...
The goal of this article is to develop a multiple-choice questions generation system that has a numb...
Open-domain question answering (QA) is an emerging information-seeking paradigm, which automatically...
Machine Translation models are trained to translate a variety of documents from one language into an...
Natural language processing (NLP) techniques had significantly improved by introducing pre-trained l...
Data annotation is critical for machine learning based natural language processing models. Although ...
The sequence-to-sequence model is a widely used model for dialogue response generators, but it tends...