Pre-trained sequence-to-sequence models have significantly improved Neural Machine Translation (NMT). Different from prior works where pre-trained models usually adopt an unidirectional decoder, this paper demonstrates that pre-training a sequence-to-sequence model but with a bidirectional decoder can produce notable performance gains for both Autoregressive and Non-autoregressive NMT. Specifically, we propose CeMAT, a conditional masked language model pre-trained on large-scale bilingual and monolingual corpora in many languages. We also introduce two simple but effective methods to enhance the CeMAT, aligned code-switching & masking and dynamic dual-masking. We conduct extensive experiments and show that our CeMAT can achieve significant ...
Neural Machine Translation (NMT) is a new approach to machine translation that has made great progre...
Without real bilingual corpus available, unsupervised Neural Machine Translation (NMT) typically req...
With the advent of deep neural networks in recent years, Neural Machine Translation (NMT) systems ha...
Unsupervised cross-lingual pretraining has achieved strong results in neural machine translation (NM...
Transformer-based autoregressive (AR) methods have achieved appealing performance for varied sequenc...
GPT-2 and BERT demonstrate the effectiveness of using pre-trained language models (LMs) on various n...
Humans benefit from communication but suffer from language barriers. Machine translation (MT) aims t...
Pre-training and fine-tuning have become the de facto paradigm in many natural language processing (...
This paper introduces a new data augmentation method for neural machine translation that can enforce...
Differently from the traditional statistical MT that decomposes the translation task into distinct s...
Monolingual data have been demonstrated to be helpful in improving translation quality of both stati...
We explore the application of neural language models to machine translation. We develop a new model ...
We explore the application of neural language models to machine translation. We develop a new model ...
Pre-training and fine-tuning have achieved great success in natural language process field. The stan...
Neural Machine Translation (NMT) typically leverages monolingual data in training through backtransl...
Neural Machine Translation (NMT) is a new approach to machine translation that has made great progre...
Without real bilingual corpus available, unsupervised Neural Machine Translation (NMT) typically req...
With the advent of deep neural networks in recent years, Neural Machine Translation (NMT) systems ha...
Unsupervised cross-lingual pretraining has achieved strong results in neural machine translation (NM...
Transformer-based autoregressive (AR) methods have achieved appealing performance for varied sequenc...
GPT-2 and BERT demonstrate the effectiveness of using pre-trained language models (LMs) on various n...
Humans benefit from communication but suffer from language barriers. Machine translation (MT) aims t...
Pre-training and fine-tuning have become the de facto paradigm in many natural language processing (...
This paper introduces a new data augmentation method for neural machine translation that can enforce...
Differently from the traditional statistical MT that decomposes the translation task into distinct s...
Monolingual data have been demonstrated to be helpful in improving translation quality of both stati...
We explore the application of neural language models to machine translation. We develop a new model ...
We explore the application of neural language models to machine translation. We develop a new model ...
Pre-training and fine-tuning have achieved great success in natural language process field. The stan...
Neural Machine Translation (NMT) typically leverages monolingual data in training through backtransl...
Neural Machine Translation (NMT) is a new approach to machine translation that has made great progre...
Without real bilingual corpus available, unsupervised Neural Machine Translation (NMT) typically req...
With the advent of deep neural networks in recent years, Neural Machine Translation (NMT) systems ha...