In domains with insufficient matched training data, language models are often constructed by interpolating component models trained from partially matched corpora. Since the n-grams from such corpora may not be of equal relevance to the target domain, we propose an n-gram weighting technique to adjust the component n-gram probabilities based on fea-tures derived from readily available segmen-tation and metadata information for each cor-pus. Using a log-linear combination of such features, the resulting model achieves up to a 1.2 % absolute word error rate reduction over a linearly interpolated baseline language model on a lecture transcription task.
We introduce a novel approach for building language models based on a systematic, recursive explorat...
Data sparsity is a large problem in natural language processing that refers to the fact that languag...
Abstract We present a modification of the traditional n-gram language modeling approach that departs...
In domains with insufficient matched training data, language models are often constructed by interpo...
International audienceThis paper describes an extension of the n-gram language model: the similar n-...
© 2015 Lyan Verwimp, Joris Pelemans, Hugo Van hamme, Patrick Wambacq. The subject of this paper is t...
This paper presents two techniques for language model (LM) adaptation. The first aims to build a mor...
In this paper, an extension of n-grams is proposed. In this extension, the memory of the model (n) i...
Statistical n-gram language modeling is used in many domains like speech recognition, language ident...
Recent progress in variable n-gram language modeling provides an efficient representation of n-gram ...
We present a tutorial introduction to n-gram models for language modeling and survey the most widely...
The recent availability of large corpora for training N-gram language models has shown the utility o...
Verwimp L., Pelemans J., Van hamme H., Wambacq P., ''Extending n-gram language models based on equiv...
In this paper, an extension of n-grams, called x-grams, is proposed. In this extension, the memory o...
In this paper, an extension of n-grams, called x-grams, is proposed. In this extension, the memory o...
We introduce a novel approach for building language models based on a systematic, recursive explorat...
Data sparsity is a large problem in natural language processing that refers to the fact that languag...
Abstract We present a modification of the traditional n-gram language modeling approach that departs...
In domains with insufficient matched training data, language models are often constructed by interpo...
International audienceThis paper describes an extension of the n-gram language model: the similar n-...
© 2015 Lyan Verwimp, Joris Pelemans, Hugo Van hamme, Patrick Wambacq. The subject of this paper is t...
This paper presents two techniques for language model (LM) adaptation. The first aims to build a mor...
In this paper, an extension of n-grams is proposed. In this extension, the memory of the model (n) i...
Statistical n-gram language modeling is used in many domains like speech recognition, language ident...
Recent progress in variable n-gram language modeling provides an efficient representation of n-gram ...
We present a tutorial introduction to n-gram models for language modeling and survey the most widely...
The recent availability of large corpora for training N-gram language models has shown the utility o...
Verwimp L., Pelemans J., Van hamme H., Wambacq P., ''Extending n-gram language models based on equiv...
In this paper, an extension of n-grams, called x-grams, is proposed. In this extension, the memory o...
In this paper, an extension of n-grams, called x-grams, is proposed. In this extension, the memory o...
We introduce a novel approach for building language models based on a systematic, recursive explorat...
Data sparsity is a large problem in natural language processing that refers to the fact that languag...
Abstract We present a modification of the traditional n-gram language modeling approach that departs...