Recent advances in NLP are brought by a range of large-scale pretrained language models (PLMs). These PLMs have brought significant performance gains for a range of NLP tasks, circumventing the need to customize complex designs for specific tasks. However, most current work focus on finetuning PLMs on a domain-specific datasets, ignoring the fact that the domain gap can lead to overfitting and even performance drop. Therefore, it is practically important to find an appropriate method to effectively adapt PLMs to a target domain of interest. Recently, a range of methods have been proposed to achieve this purpose. Early surveys on domain adaptation are not suitable for PLMs due to the sophisticated behavior exhibited by PLMs from traditional ...
Pre-trained language models (LMs) are shown to easily generate toxic language. In this work, we syst...
Domain language model (LM) adaptation consists in re-estimating probabilities of a baseline LM to be...
Domain adaptation for machine translation (MT) can be achieved by selecting training instances close...
The performance of a machine learning model trained on labeled data of a (source) domain degrades se...
When a neural language model (LM) is adapted to perform a new task, what aspects of the task predict...
With the fast growth of the amount of digitalized texts in recent years, text information management...
Large language models have transformed the field of natural language processing (NLP). Their improve...
Neural network training has been shown to be advantageous in many natural language processing appli...
Deep models must learn robust and transferable representations in order to perform well on new domai...
Recent work has demonstrated that pre-training in-domain language models can boost performance when ...
Pretrained language models (PLMs) are today the primary model for natural language processing. Despi...
The reusability of state-of-the-art Pre-trained Language Models (PLMs) is often limited by their gen...
Automatic speech recognition models are often adapted to improve their accuracy in a new domain. A p...
Pretrained language models (PTLMs) are typically learned over a large, static corpus and further fin...
The remarkable success of large language models has been driven by dense models trained on massive u...
Pre-trained language models (LMs) are shown to easily generate toxic language. In this work, we syst...
Domain language model (LM) adaptation consists in re-estimating probabilities of a baseline LM to be...
Domain adaptation for machine translation (MT) can be achieved by selecting training instances close...
The performance of a machine learning model trained on labeled data of a (source) domain degrades se...
When a neural language model (LM) is adapted to perform a new task, what aspects of the task predict...
With the fast growth of the amount of digitalized texts in recent years, text information management...
Large language models have transformed the field of natural language processing (NLP). Their improve...
Neural network training has been shown to be advantageous in many natural language processing appli...
Deep models must learn robust and transferable representations in order to perform well on new domai...
Recent work has demonstrated that pre-training in-domain language models can boost performance when ...
Pretrained language models (PLMs) are today the primary model for natural language processing. Despi...
The reusability of state-of-the-art Pre-trained Language Models (PLMs) is often limited by their gen...
Automatic speech recognition models are often adapted to improve their accuracy in a new domain. A p...
Pretrained language models (PTLMs) are typically learned over a large, static corpus and further fin...
The remarkable success of large language models has been driven by dense models trained on massive u...
Pre-trained language models (LMs) are shown to easily generate toxic language. In this work, we syst...
Domain language model (LM) adaptation consists in re-estimating probabilities of a baseline LM to be...
Domain adaptation for machine translation (MT) can be achieved by selecting training instances close...