Compared to the traditional machine reading comprehension (MRC) with limitation to the information in a passage, knowledge-driven MRC tasks aim to enable models to answer the question according to text and related commonsense knowledge. Although pre-trained Transformer-based language models (TrLMs) such as BERT and Roberta, have shown powerful perfor-mance in MRC, external knowledge such as unspoken commonsense and world knowledge still can not be used and explained explicitly. In this work, we present three simple yet effective injection methods integrated into the structure of TrLMs to fine-tune downstream knowledge-driven MRC tasks with off-the-shelf commonsense representations. Moreover, we introduce a mask mechanism for a token-level mult...
Recent days have witnessed a diverse set of knowledge injection models for pre-trained language mode...
We present Knowledge Enhanced Multimodal BART (KM-BART), which is a Transformer-based sequence-to-se...
Large-scale pre-trained language models have demonstrated strong knowledge representation ability. H...
Machine learning has a wide variety of applications in the field of natural language processing (NLP...
Contextualized representations trained over large raw text data have given remarkable improvements f...
It remains an open question whether incorporating external knowledge benefits commonsense reasoning ...
Starting from the COMET methodology by Bosselut et al. (2019), generating commonsense knowledge dire...
In this work we present a Mixture of Task-Aware Experts Network for Machine Reading Comprehension on...
Previous studies have shown the efficacy of knowledge augmentation methods in pretrained language mo...
The common practice for training commonsense models has gone from-human-to-corpus-to-machine: humans...
SemEval Task 4 Commonsense Validation and Explanation Challenge is to validate whether a system can ...
We present Pre-trained Machine Reader (PMR), a novel method to retrofit Pre-trained Language Models ...
Thesis (Ph.D.)--University of Washington, 2020For machines to understand language, they must intuiti...
Large-scale commonsense knowledge bases empower a broad range of AI applications, where the automati...
While commonsense knowledge acquisition and reasoning has traditionally been a core research topic i...
Recent days have witnessed a diverse set of knowledge injection models for pre-trained language mode...
We present Knowledge Enhanced Multimodal BART (KM-BART), which is a Transformer-based sequence-to-se...
Large-scale pre-trained language models have demonstrated strong knowledge representation ability. H...
Machine learning has a wide variety of applications in the field of natural language processing (NLP...
Contextualized representations trained over large raw text data have given remarkable improvements f...
It remains an open question whether incorporating external knowledge benefits commonsense reasoning ...
Starting from the COMET methodology by Bosselut et al. (2019), generating commonsense knowledge dire...
In this work we present a Mixture of Task-Aware Experts Network for Machine Reading Comprehension on...
Previous studies have shown the efficacy of knowledge augmentation methods in pretrained language mo...
The common practice for training commonsense models has gone from-human-to-corpus-to-machine: humans...
SemEval Task 4 Commonsense Validation and Explanation Challenge is to validate whether a system can ...
We present Pre-trained Machine Reader (PMR), a novel method to retrofit Pre-trained Language Models ...
Thesis (Ph.D.)--University of Washington, 2020For machines to understand language, they must intuiti...
Large-scale commonsense knowledge bases empower a broad range of AI applications, where the automati...
While commonsense knowledge acquisition and reasoning has traditionally been a core research topic i...
Recent days have witnessed a diverse set of knowledge injection models for pre-trained language mode...
We present Knowledge Enhanced Multimodal BART (KM-BART), which is a Transformer-based sequence-to-se...
Large-scale pre-trained language models have demonstrated strong knowledge representation ability. H...