The remarkable success of large language models has been driven by dense models trained on massive unlabeled, unstructured corpora. These corpora typically contain text from diverse, heterogeneous sources, but information about the source of the text is rarely used during training. Transferring their knowledge to a target domain is typically done by continuing training in-domain. In this paper, we introduce a method to permit domain adaptation to many diverse domains using a computationally efficient adapter approach. Our method is based on the observation that textual domains are partially overlapping, and we represent domains as a hierarchical tree structure where each node in the tree is associated with a set of adapter weights. When com...
In real-world applications, speaker recognition models often face various domain-mismatch challenges...
Deep neural networks suffer from significant performance deterioration when there exists distributio...
Pre-trained language models (LMs) are shown to easily generate toxic language. In this work, we syst...
Recent advances in NLP are brought by a range of large-scale pretrained language models (PLMs). Thes...
The performance of a machine learning model trained on labeled data of a (source) domain degrades se...
Large pretrained multilingual models, trained on dozens of languages, have delivered promising resul...
Neural network training has been shown to be advantageous in many natural language processing appli...
Automatic Speech Recognition (ASR) systems have found their use in numerous industrial applications ...
We study the highly practical but comparatively under-studied problem of latent-domain adaptation, w...
International audienceSupervised machine translation works well when the train and test data are sam...
With the fast growth of the amount of digitalized texts in recent years, text information management...
Automatic speech recognition models are often adapted to improve their accuracy in a new domain. A p...
Sentence embeddings enable us to capture the semantic similarity of short texts. Most sentence embed...
Scaling language models with more data, compute and parameters has driven significant progress in na...
In Machine Learning, a good model is one that generalizes from training data and makes accurate pred...
In real-world applications, speaker recognition models often face various domain-mismatch challenges...
Deep neural networks suffer from significant performance deterioration when there exists distributio...
Pre-trained language models (LMs) are shown to easily generate toxic language. In this work, we syst...
Recent advances in NLP are brought by a range of large-scale pretrained language models (PLMs). Thes...
The performance of a machine learning model trained on labeled data of a (source) domain degrades se...
Large pretrained multilingual models, trained on dozens of languages, have delivered promising resul...
Neural network training has been shown to be advantageous in many natural language processing appli...
Automatic Speech Recognition (ASR) systems have found their use in numerous industrial applications ...
We study the highly practical but comparatively under-studied problem of latent-domain adaptation, w...
International audienceSupervised machine translation works well when the train and test data are sam...
With the fast growth of the amount of digitalized texts in recent years, text information management...
Automatic speech recognition models are often adapted to improve their accuracy in a new domain. A p...
Sentence embeddings enable us to capture the semantic similarity of short texts. Most sentence embed...
Scaling language models with more data, compute and parameters has driven significant progress in na...
In Machine Learning, a good model is one that generalizes from training data and makes accurate pred...
In real-world applications, speaker recognition models often face various domain-mismatch challenges...
Deep neural networks suffer from significant performance deterioration when there exists distributio...
Pre-trained language models (LMs) are shown to easily generate toxic language. In this work, we syst...