In this paper we address the issue of building language models for very small training sets by adapting existing corpora. In particular we investigate methods that combine task specific unigrams with longer range models trained on a background corpus. We propose a new method to adapt class models and show how fast marginal adaptation can be improved. Instead of estimating the adaptation unigram only on the adaptation corpus, we study specific methods to adapt unigram models as well. In extensive experimental studies we show the effectiveness of the proposed methods. As compared to FMA as described in [1] we obtain an improvement of nearly 60 % for ten utterances of adaptation data. 1
Speech recognition performance is severely affected when the lexical, syntactic, or semantic charact...
Contains fulltext : 76383.pdf (author's version ) (Open Access)Workshop, 14 septem...
In this paper, we present novel techniques for performing topic adaptation on an -gram language mode...
This paper proposes a novel Language Model (LM) adaptation method based on Minimum Discrimination In...
Language modeling is critical and indispensable for many natural language ap-plications such as auto...
Building a stochastic language model (LM) for speech recog-nition requires a large corpus of target ...
This paper presents two techniques for language model (LM) adaptation. The first aims to build a mor...
In this paper, we discuss language model adaptation methods given a word list and a raw corpus. In t...
In this paper, we present a multi-layer learning approach to the language model (LM) adaptation prob...
It is today acknowledged that neural network language models outperform backoff language models in a...
Prior work shows that it is possible to expand pretrained Masked Language Models (MLMs) to new langu...
Data selection is an effective approach to domain adaptation in statistical ma-chine translation. Th...
Pre-trained language models received extensive attention in recent years. However, it is still chall...
The robust estimation of language models for new applications of spoken dialogue systems often suffe...
The performance of a machine learning model trained on labeled data of a (source) domain degrades se...
Speech recognition performance is severely affected when the lexical, syntactic, or semantic charact...
Contains fulltext : 76383.pdf (author's version ) (Open Access)Workshop, 14 septem...
In this paper, we present novel techniques for performing topic adaptation on an -gram language mode...
This paper proposes a novel Language Model (LM) adaptation method based on Minimum Discrimination In...
Language modeling is critical and indispensable for many natural language ap-plications such as auto...
Building a stochastic language model (LM) for speech recog-nition requires a large corpus of target ...
This paper presents two techniques for language model (LM) adaptation. The first aims to build a mor...
In this paper, we discuss language model adaptation methods given a word list and a raw corpus. In t...
In this paper, we present a multi-layer learning approach to the language model (LM) adaptation prob...
It is today acknowledged that neural network language models outperform backoff language models in a...
Prior work shows that it is possible to expand pretrained Masked Language Models (MLMs) to new langu...
Data selection is an effective approach to domain adaptation in statistical ma-chine translation. Th...
Pre-trained language models received extensive attention in recent years. However, it is still chall...
The robust estimation of language models for new applications of spoken dialogue systems often suffe...
The performance of a machine learning model trained on labeled data of a (source) domain degrades se...
Speech recognition performance is severely affected when the lexical, syntactic, or semantic charact...
Contains fulltext : 76383.pdf (author's version ) (Open Access)Workshop, 14 septem...
In this paper, we present novel techniques for performing topic adaptation on an -gram language mode...