We explore the suitability of self-attention models for character-level neural machine translation. We test the standard transformer model, as well as a novel variant in which the encoder block combines information from nearby characters using convolutions. We perform extensive experiments on WMT and UN datasets, testing both bilingual and multilingual translation to English using up to three input languages (French, Spanish, and Chinese). Our transformer variant consistently outperforms the standard transformer at the character-level and converges faster while learning more robust character-level alignments
Since many languages originated from a common ancestral language and influence each other, there wou...
Though early successes of Statistical Machine Translation (SMT) systems are attributed in part to th...
The University of Cambridge submission to the WMT18 news translation task focuses on the combination...
We explore the suitability of self-attention models for character-level neural machine translation. ...
Neural machine translation has been lately established as the new state of the art in machine transl...
Machine translation has received significant attention in the field of natural language processing n...
Transformer-based models have brought a radical change to neural machine translation. A key feature ...
Transformer is a neural machine translation model which revolutionizes machine translation. Compared...
Humans benefit from communication but suffer from language barriers. Machine translation (MT) aims t...
Most existing machine translation systems operate at the level of words, relying on explicit segment...
Recent work has shown that deeper character-based neural machine translation (NMT) models can outper...
International audienceRecent studies on the analysis of the multilingual representations focus on id...
Transformers have achieved state-of-the-art results across multiple NLP tasks. However, the self-att...
Neural machine translation (NMT), where neural networks are used to generate translations, has revol...
The integration of syntactic structures into Transformer machine translation has shown positive resu...
Since many languages originated from a common ancestral language and influence each other, there wou...
Though early successes of Statistical Machine Translation (SMT) systems are attributed in part to th...
The University of Cambridge submission to the WMT18 news translation task focuses on the combination...
We explore the suitability of self-attention models for character-level neural machine translation. ...
Neural machine translation has been lately established as the new state of the art in machine transl...
Machine translation has received significant attention in the field of natural language processing n...
Transformer-based models have brought a radical change to neural machine translation. A key feature ...
Transformer is a neural machine translation model which revolutionizes machine translation. Compared...
Humans benefit from communication but suffer from language barriers. Machine translation (MT) aims t...
Most existing machine translation systems operate at the level of words, relying on explicit segment...
Recent work has shown that deeper character-based neural machine translation (NMT) models can outper...
International audienceRecent studies on the analysis of the multilingual representations focus on id...
Transformers have achieved state-of-the-art results across multiple NLP tasks. However, the self-att...
Neural machine translation (NMT), where neural networks are used to generate translations, has revol...
The integration of syntactic structures into Transformer machine translation has shown positive resu...
Since many languages originated from a common ancestral language and influence each other, there wou...
Though early successes of Statistical Machine Translation (SMT) systems are attributed in part to th...
The University of Cambridge submission to the WMT18 news translation task focuses on the combination...