Although Neural Machine Translation (NMT) has achieved remarkable progress in the past several years, most NMT systems still suffer from a fundamental shortcoming as in other sequence generation tasks: errors made early in generation process are fed as inputs to the model and can be quickly amplified, harming subsequent sequence generation. To address this issue, we propose a novel model regularization method for NMT training, which aims to improve the agreement between translations generated by left-to-right (L2R) and right-to-left (R2L) NMT decoders. This goal is achieved by introducing two Kullback-Leibler divergence regularization terms into the NMT training objective to reduce the mismatch between output probabilities of L2R and R2L mo...
This work presents two different trans-lation models using recurrent neural net-works. The first one...
Although Neural Machine Translation (NMT) models have advanced state-of-the-art performance in machi...
Pre-training and fine-tuning have become the de facto paradigm in many natural language processing (...
As a new neural machine translation approach, NonAutoregressive machine Translation (NAT) has attrac...
Though early successes of Statistical Machine Translation (SMT) systems are attributed in part to th...
Neural machine translation (NMT) conducts end-to-end translation with a source language encoder and ...
Without real bilingual corpus available, unsupervised Neural Machine Translation (NMT) typically req...
Neural machine translation (NMT) heavily relies on parallel bilingual data for training. Since large...
Neural Machine Translation (NMT) is a new approach to machine translation that has made great progre...
The dominant neural machine translation (NMT) models apply unified attentional encoder-decoder neura...
UNMT tackles translation on monolingual corpora in two required languages. Since there is no explici...
An important concern in training multilingual neural machine translation (NMT) is to translate betwe...
Monolingual data have been demonstrated to be helpful in improving translation quality of both stati...
Differently from the traditional statistical MT that decomposes the translation task into distinct s...
Neural machine translation (NMT) has achieved remarkable success in producing high-quality translati...
This work presents two different trans-lation models using recurrent neural net-works. The first one...
Although Neural Machine Translation (NMT) models have advanced state-of-the-art performance in machi...
Pre-training and fine-tuning have become the de facto paradigm in many natural language processing (...
As a new neural machine translation approach, NonAutoregressive machine Translation (NAT) has attrac...
Though early successes of Statistical Machine Translation (SMT) systems are attributed in part to th...
Neural machine translation (NMT) conducts end-to-end translation with a source language encoder and ...
Without real bilingual corpus available, unsupervised Neural Machine Translation (NMT) typically req...
Neural machine translation (NMT) heavily relies on parallel bilingual data for training. Since large...
Neural Machine Translation (NMT) is a new approach to machine translation that has made great progre...
The dominant neural machine translation (NMT) models apply unified attentional encoder-decoder neura...
UNMT tackles translation on monolingual corpora in two required languages. Since there is no explici...
An important concern in training multilingual neural machine translation (NMT) is to translate betwe...
Monolingual data have been demonstrated to be helpful in improving translation quality of both stati...
Differently from the traditional statistical MT that decomposes the translation task into distinct s...
Neural machine translation (NMT) has achieved remarkable success in producing high-quality translati...
This work presents two different trans-lation models using recurrent neural net-works. The first one...
Although Neural Machine Translation (NMT) models have advanced state-of-the-art performance in machi...
Pre-training and fine-tuning have become the de facto paradigm in many natural language processing (...