We propose a new hierarchical Bayesian n-gram model of natural languages. Our model makes use of a generalization of the commonly used Dirichlet distributions called Pitman-Yor processes which produce power-law distributions more closely resembling those in natural languages. We show that an approximation to the hierarchical Pitman-Yor language model recovers the exact formulation of interpolated Kneser-Ney, one of the best smoothing methods for n-gram language models. Experiments verify that our model gives cross entropy results superior to interpolated Kneser-Ney and comparable to modified Kneser-Ney. © 2006 Association for Computational Linguistics
We describe a unified probabilistic framework for statistical language modeling-the latent maximum e...
International audienceThis paper describes an extension of the n-gram language model: the similar n-...
The Hierarchical Pitman Yor Process Language Model (HPYLM) is a Bayesian language model based on a n...
In this paper we present a doubly hierarchical Pitman-Yor process language model. Its bottom layer o...
We present an approximation to the Bayesian hierarchical Pitman-Yor process language model which mai...
Standard statistical models of language fail to capture one of the most striking properties of natur...
In this paper we investigate the application of a novel technique for language modeling - a hierarch...
In this work we address the challenge of augmenting n-gram language models according to prior lingui...
Traditional n-gram language models are widely used in state-of-the-art large vocabulary speech recog...
We present a nonparametric Bayesian model of tree structures based on the hierarchical Dirichlet pro...
We introduce a novel approach for build-ing language models based on a system-atic, recursive explor...
We introduce a novel approach for building language models based on a systematic, recursive explorat...
In language modeling, it is nearly always assumed that documents are generated by sampling from a mu...
In this work we address the problem of unsupervised part-of-speech induction by bringing together se...
In this paper, we propose a new language model based on depen-dent word sequences organized in a mul...
We describe a unified probabilistic framework for statistical language modeling-the latent maximum e...
International audienceThis paper describes an extension of the n-gram language model: the similar n-...
The Hierarchical Pitman Yor Process Language Model (HPYLM) is a Bayesian language model based on a n...
In this paper we present a doubly hierarchical Pitman-Yor process language model. Its bottom layer o...
We present an approximation to the Bayesian hierarchical Pitman-Yor process language model which mai...
Standard statistical models of language fail to capture one of the most striking properties of natur...
In this paper we investigate the application of a novel technique for language modeling - a hierarch...
In this work we address the challenge of augmenting n-gram language models according to prior lingui...
Traditional n-gram language models are widely used in state-of-the-art large vocabulary speech recog...
We present a nonparametric Bayesian model of tree structures based on the hierarchical Dirichlet pro...
We introduce a novel approach for build-ing language models based on a system-atic, recursive explor...
We introduce a novel approach for building language models based on a systematic, recursive explorat...
In language modeling, it is nearly always assumed that documents are generated by sampling from a mu...
In this work we address the problem of unsupervised part-of-speech induction by bringing together se...
In this paper, we propose a new language model based on depen-dent word sequences organized in a mul...
We describe a unified probabilistic framework for statistical language modeling-the latent maximum e...
International audienceThis paper describes an extension of the n-gram language model: the similar n-...
The Hierarchical Pitman Yor Process Language Model (HPYLM) is a Bayesian language model based on a n...