Inference tasks such as answer sentence selection (AS2) or fact verification are typically solved by fine-tuning transformer-based models as individual sentence-pair classifiers. Recent studies show that these tasks benefit from modeling dependencies across multiple candidate sentences jointly. In this paper, we first show that popular pre-trained transformers perform poorly when used for fine-tuning on multi-candidate inference tasks. We then propose a new pre-training objective that models the paragraph-level semantics across multiple input sentences. Our evaluation on three AS2 and one fact verification datasets demonstrates the superiority of our pre-training technique over the traditional ones for transformers used as joint models for ...
Language Generation Models produce words based on the previous context. Although existing methods of...
© 2019 Association for Computational Linguistics The task of Natural Language Inference (NLI) is wid...
Large transformer models can highly improve Answer Sentence Selection (AS2) tasks, but their high co...
An important task for designing QA systems is answer sentence selection (AS2): selecting the sentenc...
We propose TandA, an effective technique for fine-tuning pre-trained Transformer models for natural ...
Recent prompt-based approaches allow pretrained language models to achieve strong performances on fe...
Structure prediction (SP) tasks are important in natural language understanding in the sense that th...
The ability to reason with natural language is a fundamental prerequisite for many NLP tasks such as...
The thesis describes the creation of an English dataset built to analyze the inference relation betw...
Recent studies on knowledge graphs (KGs) show that path-based methods empowered by pre-trained langu...
International audienceWe probe pre-trained transformer language models for bridging inference. We fi...
Transformer models have achieved promising results on natural language processing (NLP) tasks includ...
While Pre-trained Language Models (PLMs) internalize a great amount of world knowledge, they have be...
Sentence Split and Rephrase aims to break down a complex sentence into several simple sentences with...
Pre-trained Language Models recently gained traction in the Natural Language Processing (NLP) domain...
Language Generation Models produce words based on the previous context. Although existing methods of...
© 2019 Association for Computational Linguistics The task of Natural Language Inference (NLI) is wid...
Large transformer models can highly improve Answer Sentence Selection (AS2) tasks, but their high co...
An important task for designing QA systems is answer sentence selection (AS2): selecting the sentenc...
We propose TandA, an effective technique for fine-tuning pre-trained Transformer models for natural ...
Recent prompt-based approaches allow pretrained language models to achieve strong performances on fe...
Structure prediction (SP) tasks are important in natural language understanding in the sense that th...
The ability to reason with natural language is a fundamental prerequisite for many NLP tasks such as...
The thesis describes the creation of an English dataset built to analyze the inference relation betw...
Recent studies on knowledge graphs (KGs) show that path-based methods empowered by pre-trained langu...
International audienceWe probe pre-trained transformer language models for bridging inference. We fi...
Transformer models have achieved promising results on natural language processing (NLP) tasks includ...
While Pre-trained Language Models (PLMs) internalize a great amount of world knowledge, they have be...
Sentence Split and Rephrase aims to break down a complex sentence into several simple sentences with...
Pre-trained Language Models recently gained traction in the Natural Language Processing (NLP) domain...
Language Generation Models produce words based on the previous context. Although existing methods of...
© 2019 Association for Computational Linguistics The task of Natural Language Inference (NLI) is wid...
Large transformer models can highly improve Answer Sentence Selection (AS2) tasks, but their high co...