This paper considers continual learning of large-scale pretrained neural machine translation model without accessing the previous training data or introducing model separation. We argue that the widely used regularization-based methods, which perform multi-objective learning with an auxiliary loss, suffer from the misestimate problem and cannot always achieve a good balance between the previous and new tasks. To solve the problem, we propose a two-stage training method based on the local features of the real loss. We first search low forgetting risk regions, where the model can retain the performance on the previous task as the parameters are updated, to avoid the catastrophic forgetting problem. Then we can continually train the model with...
Catastrophic forgetting (CF) happens whenever a neural network overwrites past knowledge while being...
Pretrained language models (PLMs) are today the primary model for natural language processing. Despi...
Continual learning is a framework of learning in which we aim to move beyond the limitations of stan...
Recent literature has demonstrated the potential of multilingual Neural Machine Translation (mNMT) m...
GPT-2 and BERT demonstrate the effectiveness of using pre-trained language models (LMs) on various n...
The lifelong learning paradigm in machine learning is an attractive alternative to the more prominen...
Multilingual speech recognition with neural networks is often implemented with batch-learning, when ...
Neural machine translation (NMT), where neural networks are used to generate translations, has revol...
Kenneweg P, Schulz A, Schroeder S, Hammer B. Intelligent Learning Rate Distribution to Reduce Catast...
Pre-trained models are nowadays a fundamental component of machine learning research. In continual l...
With the advent of deep neural networks in recent years, Neural Machine Translation (NMT) systems ha...
Improving machine translation (MT) by learning from human post-edits is a powerful solution that is ...
Neural machine translation is known to require large numbers of parallel training sentences, which g...
The Contrastive Language-Image Pre-training (CLIP) Model is a recently proposed large-scale pre-trai...
We analyze the learning dynamics of neural language and translation models using Loss Change Allocat...
Catastrophic forgetting (CF) happens whenever a neural network overwrites past knowledge while being...
Pretrained language models (PLMs) are today the primary model for natural language processing. Despi...
Continual learning is a framework of learning in which we aim to move beyond the limitations of stan...
Recent literature has demonstrated the potential of multilingual Neural Machine Translation (mNMT) m...
GPT-2 and BERT demonstrate the effectiveness of using pre-trained language models (LMs) on various n...
The lifelong learning paradigm in machine learning is an attractive alternative to the more prominen...
Multilingual speech recognition with neural networks is often implemented with batch-learning, when ...
Neural machine translation (NMT), where neural networks are used to generate translations, has revol...
Kenneweg P, Schulz A, Schroeder S, Hammer B. Intelligent Learning Rate Distribution to Reduce Catast...
Pre-trained models are nowadays a fundamental component of machine learning research. In continual l...
With the advent of deep neural networks in recent years, Neural Machine Translation (NMT) systems ha...
Improving machine translation (MT) by learning from human post-edits is a powerful solution that is ...
Neural machine translation is known to require large numbers of parallel training sentences, which g...
The Contrastive Language-Image Pre-training (CLIP) Model is a recently proposed large-scale pre-trai...
We analyze the learning dynamics of neural language and translation models using Loss Change Allocat...
Catastrophic forgetting (CF) happens whenever a neural network overwrites past knowledge while being...
Pretrained language models (PLMs) are today the primary model for natural language processing. Despi...
Continual learning is a framework of learning in which we aim to move beyond the limitations of stan...