Machine learning has a wide variety of applications in the field of natural language processing (NLP). One such application is fine-tuning large pre-trained models to a wide variety of tasks. In this work, we propose methods to enhance these large language models by infusing them with information found in commonsense knowledge bases. Commonsense is basic knowledge about the world that humans are expected to have and is needed to achieve efficient communication. Often times, to understand texts, a person must use their commonsense to make implicit inferences based on what is explicitly presented in text. We harness the power of relational graph convolutional networks (RGCNs) to encode meaningful commonsense information from graphs and introd...
Compared to the traditional machine reading comprehension (MRC) with limitation to the information i...
Intelligent systems are expected to make smart human-like decisions based on accumulated commonsense...
Combining structured information with language models is a standing problem in NLP. Building on prev...
Contextualized representations trained over large raw text data have given remarkable improvements f...
Thesis (Ph.D.)--University of Washington, 2020For machines to understand language, they must intuiti...
Thesis (Ph.D.)--University of Washington, 2021Along with the meteoric rise of computation-hungry mod...
Commonsense reasoning is an important aspect of building robust AI systems and is receiving signific...
Acquiring commonsense knowledge and reasoning is an important goal in modern NLP research. Despite m...
Recent years have brought about a renewed interest in commonsense representation and reasoning in th...
Automatic KB completion for commonsense knowledge graphs (e.g., ATOMIC and ConceptNet) poses unique ...
Incorporating prior knowledge can improve existing pre-training models in cloze-style machine readin...
Modern language models are strong at generating grammatically correct, natural lan- guage. However, ...
Starting from the COMET methodology by Bosselut et al. (2019), generating commonsense knowledge dire...
Large-scale pre-trained language models have demonstrated strong knowledge representation ability. H...
Generative commonsense reasoning which aims to empower machines to generate sentences with the capac...
Compared to the traditional machine reading comprehension (MRC) with limitation to the information i...
Intelligent systems are expected to make smart human-like decisions based on accumulated commonsense...
Combining structured information with language models is a standing problem in NLP. Building on prev...
Contextualized representations trained over large raw text data have given remarkable improvements f...
Thesis (Ph.D.)--University of Washington, 2020For machines to understand language, they must intuiti...
Thesis (Ph.D.)--University of Washington, 2021Along with the meteoric rise of computation-hungry mod...
Commonsense reasoning is an important aspect of building robust AI systems and is receiving signific...
Acquiring commonsense knowledge and reasoning is an important goal in modern NLP research. Despite m...
Recent years have brought about a renewed interest in commonsense representation and reasoning in th...
Automatic KB completion for commonsense knowledge graphs (e.g., ATOMIC and ConceptNet) poses unique ...
Incorporating prior knowledge can improve existing pre-training models in cloze-style machine readin...
Modern language models are strong at generating grammatically correct, natural lan- guage. However, ...
Starting from the COMET methodology by Bosselut et al. (2019), generating commonsense knowledge dire...
Large-scale pre-trained language models have demonstrated strong knowledge representation ability. H...
Generative commonsense reasoning which aims to empower machines to generate sentences with the capac...
Compared to the traditional machine reading comprehension (MRC) with limitation to the information i...
Intelligent systems are expected to make smart human-like decisions based on accumulated commonsense...
Combining structured information with language models is a standing problem in NLP. Building on prev...