The integration of syntactic structures into Transformer machine translation has shown positive results, but to our knowledge, no work has attempted to do so with semantic structures. In this work we propose two novel parameter-free methods for injecting semantic information into Transformers, both rely on semantics-aware masking of (some of) the attention heads. One such method operates on the encoder, through a Scene-Aware Self-Attention (SASA) head. Another on the decoder, through a Scene-Aware Cross-Attention (SACrA) head. We show a consistent improvement over the vanilla Transformer and syntax-aware models for four language pairs. We further show an additional gain when using both semantic and syntactic structures in some language pair...
Humans benefit from communication but suffer from language barriers. Machine translation (MT) aims t...
Scene text recognition (STR) enables computers to recognize and read the text in various real-world ...
In this paper we incorporate semantic supersensetags and syntactic supertag features into EN–FR and ...
Neural machine translation has been lately established as the new state of the art in machine transl...
Transformer-based models have brought a radical change to neural machine translation. A key feature ...
Introducing factors such as linguistic features has long been proposed in machine translation to imp...
The utility of linguistic annotation in neural machine translation seemed to had been established in...
We explore the suitability of self-attention models for character-level neural machine translation. ...
In this thesis, I explore neural machine translation (NMT) models via targeted investigation of vari...
Neural machine translation has considerably improved the quality of automatic translations by learni...
In Neural Machine Translation (NMT), each token prediction is conditioned on the source sentence and...
Machine translation has received significant attention in the field of natural language processing n...
Natural language processing (NLP) techniques had significantly improved by introducing pre-trained l...
Document-level Neural Machine Translation (DocNMT) has been proven crucial for handling discourse ph...
Syntax knowledge contributes its powerful strength in Neural machine translation (NMT) tasks. Early ...
Humans benefit from communication but suffer from language barriers. Machine translation (MT) aims t...
Scene text recognition (STR) enables computers to recognize and read the text in various real-world ...
In this paper we incorporate semantic supersensetags and syntactic supertag features into EN–FR and ...
Neural machine translation has been lately established as the new state of the art in machine transl...
Transformer-based models have brought a radical change to neural machine translation. A key feature ...
Introducing factors such as linguistic features has long been proposed in machine translation to imp...
The utility of linguistic annotation in neural machine translation seemed to had been established in...
We explore the suitability of self-attention models for character-level neural machine translation. ...
In this thesis, I explore neural machine translation (NMT) models via targeted investigation of vari...
Neural machine translation has considerably improved the quality of automatic translations by learni...
In Neural Machine Translation (NMT), each token prediction is conditioned on the source sentence and...
Machine translation has received significant attention in the field of natural language processing n...
Natural language processing (NLP) techniques had significantly improved by introducing pre-trained l...
Document-level Neural Machine Translation (DocNMT) has been proven crucial for handling discourse ph...
Syntax knowledge contributes its powerful strength in Neural machine translation (NMT) tasks. Early ...
Humans benefit from communication but suffer from language barriers. Machine translation (MT) aims t...
Scene text recognition (STR) enables computers to recognize and read the text in various real-world ...
In this paper we incorporate semantic supersensetags and syntactic supertag features into EN–FR and ...