In this paper, a new n-gram language model compression method is proposed for applications in handheld devices, such as mobiles, PDAs, and handheld PCs. Compared with the traditional methods, the use of the proposed method can compress the model to a great extent with good performance preserved. The proposed method includes three aspects. The language model parameters are detailedly analyzed and a criterion based on the probability and the importance of n-grams is used to determine which n-grams should be kept and which be removed. A curving compressing function is proposed to be used to compress the n-gram count values in the full language model. And a code table is extracted and used to estimate the probabilities of bi-grams. Our experime...
In this paper an approach is proposed to combine different order N-grams based on the discriminative...
We present a tutorial introduction to n-gram models for language modeling and survey the most widely...
Research in speech recognition and machine translation is boosting the use of large scale n-gram lan...
In this paper, a new n-gram language model compression method is proposed for applications in handhe...
This paper describes a novel approach of compressing large trigram language models, which uses scala...
This paper describes two techniques for reducing the size of statistical back-off-gram language mode...
ICSLP1998: the 5th International Conference on Spoken Language Processing, November 30 - December 4...
In this paper, an extension of n-grams, called x-grams, is proposed. In this extension, the memory o...
In this paper, an extension of n-grams is proposed. In this extension, the memory of the model (n) i...
N-gram language models are an essential component in statistical natural language processing systems...
Statistical n-gram language modeling is used in many domains like speech recognition, language ident...
The best general-purpose compression schemes make their gains by estimating a probability distributi...
Language model in Natural Language Processing is one of the most important fields carried out in the...
In domains with insufficient matched training data, language models are often constructed by interpo...
It has been a matter of convenience that language codes have utilized statistical aspects of easily ...
In this paper an approach is proposed to combine different order N-grams based on the discriminative...
We present a tutorial introduction to n-gram models for language modeling and survey the most widely...
Research in speech recognition and machine translation is boosting the use of large scale n-gram lan...
In this paper, a new n-gram language model compression method is proposed for applications in handhe...
This paper describes a novel approach of compressing large trigram language models, which uses scala...
This paper describes two techniques for reducing the size of statistical back-off-gram language mode...
ICSLP1998: the 5th International Conference on Spoken Language Processing, November 30 - December 4...
In this paper, an extension of n-grams, called x-grams, is proposed. In this extension, the memory o...
In this paper, an extension of n-grams is proposed. In this extension, the memory of the model (n) i...
N-gram language models are an essential component in statistical natural language processing systems...
Statistical n-gram language modeling is used in many domains like speech recognition, language ident...
The best general-purpose compression schemes make their gains by estimating a probability distributi...
Language model in Natural Language Processing is one of the most important fields carried out in the...
In domains with insufficient matched training data, language models are often constructed by interpo...
It has been a matter of convenience that language codes have utilized statistical aspects of easily ...
In this paper an approach is proposed to combine different order N-grams based on the discriminative...
We present a tutorial introduction to n-gram models for language modeling and survey the most widely...
Research in speech recognition and machine translation is boosting the use of large scale n-gram lan...