For machine reading comprehension, the capacity of effectively modeling the linguistic knowledge from the detail-riddled and lengthy passages and getting ride of the noises is essential to improve its performance. Traditional attentive models attend to all words without explicit constraint, which results in inaccurate concentration on some dispensable words. In this work, we propose using syntax to guide the text modeling by incorporating explicit syntactic constraints into attention mechanism for better linguistically motivated word representations. In detail, for self-attention network (SAN) sponsored Transformer-based encoder, we introduce syntactic dependency of interest (SDOI) design into the SAN to form an SDOI-SAN with syntax-guided ...
Comprehending unstructured text is a challenging task for machines because it involves understanding...
Syntax — the study of the hierarchical structure of language — has long featured as a prominent rese...
Thesis: Ph. D., Massachusetts Institute of Technology, Department of Brain and Cognitive Sciences, F...
In recent years, pre-trained language models, represented by the bidirectional encoder representatio...
Syntax knowledge contributes its powerful strength in Neural machine translation (NMT) tasks. Early ...
Natural language processing (NLP) techniques had significantly improved by introducing pre-trained l...
Neural machine translation has been lately established as the new state of the art in machine transl...
The integration of syntactic structures into Transformer machine translation has shown positive resu...
Recent progress in NLP witnessed the development of large-scale pre-trained language models (GPT, BE...
Machine translation has received significant attention in the field of natural language processing n...
Deep neural models (e.g. Transformer) naturally learn spurious features, which create a ``shortcut''...
We propose a novel Transformer encoder-based architecture with syntactical knowledge encoded for int...
The core of self-supervised learning for pre-training language models includes pre-training task des...
Large pretrained language models using the transformer neural network architecture are becoming a do...
The utility of linguistic annotation in neural machine translation seemed to had been established in...
Comprehending unstructured text is a challenging task for machines because it involves understanding...
Syntax — the study of the hierarchical structure of language — has long featured as a prominent rese...
Thesis: Ph. D., Massachusetts Institute of Technology, Department of Brain and Cognitive Sciences, F...
In recent years, pre-trained language models, represented by the bidirectional encoder representatio...
Syntax knowledge contributes its powerful strength in Neural machine translation (NMT) tasks. Early ...
Natural language processing (NLP) techniques had significantly improved by introducing pre-trained l...
Neural machine translation has been lately established as the new state of the art in machine transl...
The integration of syntactic structures into Transformer machine translation has shown positive resu...
Recent progress in NLP witnessed the development of large-scale pre-trained language models (GPT, BE...
Machine translation has received significant attention in the field of natural language processing n...
Deep neural models (e.g. Transformer) naturally learn spurious features, which create a ``shortcut''...
We propose a novel Transformer encoder-based architecture with syntactical knowledge encoded for int...
The core of self-supervised learning for pre-training language models includes pre-training task des...
Large pretrained language models using the transformer neural network architecture are becoming a do...
The utility of linguistic annotation in neural machine translation seemed to had been established in...
Comprehending unstructured text is a challenging task for machines because it involves understanding...
Syntax — the study of the hierarchical structure of language — has long featured as a prominent rese...
Thesis: Ph. D., Massachusetts Institute of Technology, Department of Brain and Cognitive Sciences, F...