ICSLP1998: the 5th International Conference on Spoken Language Processing, November 30 - December 4, 1998, Sydney, Australia.In this paper we propose an algorithm for reducing the size of back-off N-gram models, with less affecting its performance than the traditional cutoff method. The algorithm is based on the Maximum Likelihood (ML) estimation and realizes an N-gram language model with a given number of N-gram probability parameters that minimize the training set perplexity. To confirm the effectiveness of our algorithm, we apply it to trigram and bigram models, and the experiments in terms of perplexity and word error rate in a dictation system are carried out
We propose a novel method for using the World Wide Web to acquire trigram estimates for statistical ...
N-gram language models are an essential component in statistical natural language processing systems...
International audienceIn this article we deal with the text segmentation problem in statistical lang...
This paper describes a novel approach of compressing large trigram language models, which uses scala...
In this paper several methods are proposed for reducing the size of a trigram language model (LM), w...
When a trigram backoff language model is created from a large body of text, trigrams and bigrams tha...
In this paper, a new n-gram language model compression method is proposed for applications in handhe...
In this paper, an extension of n-grams is proposed. In this extension, the memory of the model (n) i...
In domains with insufficient matched training data, language models are often constructed by interpo...
In this paper, an extension of n-grams, called x-grams, is proposed. In this extension, the memory o...
This paper describes two techniques for reducing the size of statistical back-off-gram language mode...
In this paper we present two new techniques for language modeling in speech recognition. The rst tec...
This paper presents two techniques for language model (LM) adaptation. The first aims to build a mor...
Previous attempts to automatically determine multi-words as the basic unit for language modeling hav...
International audienceThis paper deals with the combination of a trigram and a triclass. This combin...
We propose a novel method for using the World Wide Web to acquire trigram estimates for statistical ...
N-gram language models are an essential component in statistical natural language processing systems...
International audienceIn this article we deal with the text segmentation problem in statistical lang...
This paper describes a novel approach of compressing large trigram language models, which uses scala...
In this paper several methods are proposed for reducing the size of a trigram language model (LM), w...
When a trigram backoff language model is created from a large body of text, trigrams and bigrams tha...
In this paper, a new n-gram language model compression method is proposed for applications in handhe...
In this paper, an extension of n-grams is proposed. In this extension, the memory of the model (n) i...
In domains with insufficient matched training data, language models are often constructed by interpo...
In this paper, an extension of n-grams, called x-grams, is proposed. In this extension, the memory o...
This paper describes two techniques for reducing the size of statistical back-off-gram language mode...
In this paper we present two new techniques for language modeling in speech recognition. The rst tec...
This paper presents two techniques for language model (LM) adaptation. The first aims to build a mor...
Previous attempts to automatically determine multi-words as the basic unit for language modeling hav...
International audienceThis paper deals with the combination of a trigram and a triclass. This combin...
We propose a novel method for using the World Wide Web to acquire trigram estimates for statistical ...
N-gram language models are an essential component in statistical natural language processing systems...
International audienceIn this article we deal with the text segmentation problem in statistical lang...