Syntax knowledge contributes its powerful strength in Neural machine translation (NMT) tasks. Early NMT works supposed that syntax details can be automatically learned from numerous texts via attention networks. However, succeeding researches pointed out that limited by the uncontrolled nature of attention computation, the NMT model requires an external syntax to capture the deep syntactic awareness. Although existing syntax-aware NMT methods have born great fruits in combining syntax, the additional workloads they introduced render the model heavy and slow. Particularly, these efforts scarcely involve the Transformer-based NMT and modify its core self-attention network (SAN). To this end, we propose a parameter-free, Dependency-scaled Self...
Neural machine translation (NMT), where neural networks are used to generate translations, has revol...
Despite the progress made in sentence-level NMT, current systems still fall short at achieving fluen...
Attention-based autoregressive models have achieved state-of-the-art performance in various sequence...
Attention mechanism, including global attention and local attention, plays a key role in neural mach...
Recently, non-recurrent architectures (convolutional, self-attentional) have outperformed RNNs ...
This paper presents an extension of neural machine translation (NMT) model to incorporate additional...
Neural machine translation has been lately established as the new state of the art in machine transl...
An attentional mechanism has lately been used to improve neural machine transla-tion (NMT) by select...
The utility of linguistic annotation in neural machine translation seemed to had been established in...
The integration of syntactic structures into Transformer machine translation has shown positive resu...
Humans benefit from communication but suffer from language barriers. Machine translation (MT) aims t...
Multi-head self-attention recently attracts enormous interest owing to its specialized functions, si...
For machine reading comprehension, the capacity of effectively modeling the linguistic knowledge fro...
We explored the syntactic information encoded implicitly by neural machine translation (NMT) models ...
In this thesis, I explore neural machine translation (NMT) models via targeted investigation of vari...
Neural machine translation (NMT), where neural networks are used to generate translations, has revol...
Despite the progress made in sentence-level NMT, current systems still fall short at achieving fluen...
Attention-based autoregressive models have achieved state-of-the-art performance in various sequence...
Attention mechanism, including global attention and local attention, plays a key role in neural mach...
Recently, non-recurrent architectures (convolutional, self-attentional) have outperformed RNNs ...
This paper presents an extension of neural machine translation (NMT) model to incorporate additional...
Neural machine translation has been lately established as the new state of the art in machine transl...
An attentional mechanism has lately been used to improve neural machine transla-tion (NMT) by select...
The utility of linguistic annotation in neural machine translation seemed to had been established in...
The integration of syntactic structures into Transformer machine translation has shown positive resu...
Humans benefit from communication but suffer from language barriers. Machine translation (MT) aims t...
Multi-head self-attention recently attracts enormous interest owing to its specialized functions, si...
For machine reading comprehension, the capacity of effectively modeling the linguistic knowledge fro...
We explored the syntactic information encoded implicitly by neural machine translation (NMT) models ...
In this thesis, I explore neural machine translation (NMT) models via targeted investigation of vari...
Neural machine translation (NMT), where neural networks are used to generate translations, has revol...
Despite the progress made in sentence-level NMT, current systems still fall short at achieving fluen...
Attention-based autoregressive models have achieved state-of-the-art performance in various sequence...