Neural machine translation has been lately established as the new state of the art in machine translation, especially with the Transformer model. This model emphasized the importance of self-attention mechanism and sug- gested that it could capture some linguistic phenomena. However, this claim has not been examined thoroughly, so we propose two main groups of meth- ods to examine the relation between these two. Our methods aim to im- prove the translation performance by directly manipulating the self-attention layer. The first group focuses on enriching the encoder with source-side syn- tax with tree-related position embeddings or our novel specialized attention heads. The second group is a joint translation and parsing model leveraging se...
Transformers have achieved state-of-the-art results across multiple NLP tasks. However, the self-att...
Transformer is a neural machine translation model which revolutionizes machine translation. Compared...
Introducing factors such as linguistic features has long been proposed in machine translation to imp...
Transformer-based models have brought a radical change to neural machine translation. A key feature ...
We explore the suitability of self-attention models for character-level neural machine translation. ...
Machine translation has received significant attention in the field of natural language processing n...
The utility of linguistic annotation in neural machine translation seemed to had been established in...
The Transformer model is a very recent, fast and powerful discovery in neural machine translation. W...
Multi-head self-attention is a key component of the Transformer, a state-of-the-art architecture for...
International audienceRecent studies on the analysis of the multilingual representations focus on id...
The integration of syntactic structures into Transformer machine translation has shown positive resu...
Humans benefit from communication but suffer from language barriers. Machine translation (MT) aims t...
Though early successes of Statistical Machine Translation (SMT) systems are attributed in part to th...
In this thesis, I explore neural machine translation (NMT) models via targeted investigation of vari...
An attentional mechanism has lately been used to improve neural machine transla-tion (NMT) by select...
Transformers have achieved state-of-the-art results across multiple NLP tasks. However, the self-att...
Transformer is a neural machine translation model which revolutionizes machine translation. Compared...
Introducing factors such as linguistic features has long been proposed in machine translation to imp...
Transformer-based models have brought a radical change to neural machine translation. A key feature ...
We explore the suitability of self-attention models for character-level neural machine translation. ...
Machine translation has received significant attention in the field of natural language processing n...
The utility of linguistic annotation in neural machine translation seemed to had been established in...
The Transformer model is a very recent, fast and powerful discovery in neural machine translation. W...
Multi-head self-attention is a key component of the Transformer, a state-of-the-art architecture for...
International audienceRecent studies on the analysis of the multilingual representations focus on id...
The integration of syntactic structures into Transformer machine translation has shown positive resu...
Humans benefit from communication but suffer from language barriers. Machine translation (MT) aims t...
Though early successes of Statistical Machine Translation (SMT) systems are attributed in part to th...
In this thesis, I explore neural machine translation (NMT) models via targeted investigation of vari...
An attentional mechanism has lately been used to improve neural machine transla-tion (NMT) by select...
Transformers have achieved state-of-the-art results across multiple NLP tasks. However, the self-att...
Transformer is a neural machine translation model which revolutionizes machine translation. Compared...
Introducing factors such as linguistic features has long been proposed in machine translation to imp...