Non-autoregressive approaches aim to improve the inference speed of translation models, particularly those that generate output in a one-pass forward manner. However, these approaches often suffer from a significant drop in translation quality compared to autoregressive models. This paper introduces a series of innovative techniques to enhance the translation quality of Non-Autoregressive Translation (NAT) models while maintaining a substantial acceleration in inference speed. We propose fine-tuning Pretrained Multilingual Language Models (PMLMs) with the CTC loss to train NAT models effectively. Furthermore, we adopt the MASK insertion scheme for up-sampling instead of token duplication, and we present an embedding distillation method to f...
International audienceNon-autoregressive machine translation (NAT) has recently made great progress....
2014-07-28The goal of machine translation is to translate from one natural language into another usi...
We participated in the WMT 2016 shared news translation task on English ↔ Chinese language pair. Ou...
Non-autoregressive translation (NAT) models remove the dependence on previous target tokens and gene...
Non-autoregressive (NAR) generation, which is first proposed in neural machine translation (NMT) to ...
Non-autoregressive translation (NAT) models, which remove the dependence on previous target tokens f...
Non-autoregressive approaches aim to improve the inference speed of translation models by only requi...
Non-autoregressive neural machine translation (NAT) models are proposed to accelerate the inference ...
How do we perform efficient inference while retaining high translation quality? Existing neural mach...
Benefiting from the sequence-level knowledge distillation, the Non-Autoregressive Transformer (NAT) ...
GPT-2 and BERT demonstrate the effectiveness of using pre-trained language models (LMs) on various n...
Pre-trained sequence-to-sequence models have significantly improved Neural Machine Translation (NMT)...
Non-autoregressive machine translation (NAT) models have lower translation quality than autoregressi...
Non-Autoregressive Neural Machine Translation (NAT) achieves significant decoding speedup through ge...
Humans benefit from communication but suffer from language barriers. Machine translation (MT) aims t...
International audienceNon-autoregressive machine translation (NAT) has recently made great progress....
2014-07-28The goal of machine translation is to translate from one natural language into another usi...
We participated in the WMT 2016 shared news translation task on English ↔ Chinese language pair. Ou...
Non-autoregressive translation (NAT) models remove the dependence on previous target tokens and gene...
Non-autoregressive (NAR) generation, which is first proposed in neural machine translation (NMT) to ...
Non-autoregressive translation (NAT) models, which remove the dependence on previous target tokens f...
Non-autoregressive approaches aim to improve the inference speed of translation models by only requi...
Non-autoregressive neural machine translation (NAT) models are proposed to accelerate the inference ...
How do we perform efficient inference while retaining high translation quality? Existing neural mach...
Benefiting from the sequence-level knowledge distillation, the Non-Autoregressive Transformer (NAT) ...
GPT-2 and BERT demonstrate the effectiveness of using pre-trained language models (LMs) on various n...
Pre-trained sequence-to-sequence models have significantly improved Neural Machine Translation (NMT)...
Non-autoregressive machine translation (NAT) models have lower translation quality than autoregressi...
Non-Autoregressive Neural Machine Translation (NAT) achieves significant decoding speedup through ge...
Humans benefit from communication but suffer from language barriers. Machine translation (MT) aims t...
International audienceNon-autoregressive machine translation (NAT) has recently made great progress....
2014-07-28The goal of machine translation is to translate from one natural language into another usi...
We participated in the WMT 2016 shared news translation task on English ↔ Chinese language pair. Ou...