We present an empirical study of the applicability of Probabilistic Lexicalized Tree Insertion Grammars (PLTIG), a lexicalized counterpart to Probabilistic Context-Free Grammars (PCFG), to problems in stochastic naturallanguage processing. Comparing the performance of PLTIGs with non-hierarchical N-gram models and PCFGs, we show that PLTIG combines the best aspects of both, with language modeling capability comparable to N-grams, and improved parsing performance over its nonlexicalized counterpart. Furthermore, training of PLTIGs displays faster convergence than PCFGs. 1 Introduction There are many advantages to expressing a grammar in a lexicalized form, where an observable word of the language is encoded in each grammar rule. First, t...
The task of unsupervised induction of probabilistic context-free grammars (PCFGs) has attracted a lo...
The task of unsupervised induction of probabilistic context-free grammars (PCFGs) has attracted a lo...
Language models for speech recognition typically use a probability model of the form Pr(an/a1, a2, ....
We present an empirical study of the applicability of Probabilistic Lexicalized Tree Insertion Gramm...
We evaluated probabilistic lexicalized tree-insertion grammars (PLTIGs) on a classification task rel...
We evaluated probabilistic lexicalized tree-insertion grammars (PLTIGs) on a classification task rel...
Increasingly, researchers developing statistical machine translation systems have moved to incorpora...
Inducing a grammar from text has proven to be a notoriously challenging learning task despite decade...
Stochastic categorial grammars (SCGs) are introduced as a more appropriate formalism for statistical...
Abstract. Although state-of-the-art parsers for natural language are lexicalized, it was recently sh...
Building models of language is a central task in natural language processing. Traditionally, languag...
This paper describes a fully implemented, broad coverage model of human syntactic processing. The mo...
Although state-of-the-art parsers for natural language are lexicalized, it was recently shown that a...
We address the issue of how to associate frequency information with lexicalized grammar formalisms, ...
The task of unsupervised induction of probabilistic context-free grammars (PCFGs) has attracted a lo...
The task of unsupervised induction of probabilistic context-free grammars (PCFGs) has attracted a lo...
The task of unsupervised induction of probabilistic context-free grammars (PCFGs) has attracted a lo...
Language models for speech recognition typically use a probability model of the form Pr(an/a1, a2, ....
We present an empirical study of the applicability of Probabilistic Lexicalized Tree Insertion Gramm...
We evaluated probabilistic lexicalized tree-insertion grammars (PLTIGs) on a classification task rel...
We evaluated probabilistic lexicalized tree-insertion grammars (PLTIGs) on a classification task rel...
Increasingly, researchers developing statistical machine translation systems have moved to incorpora...
Inducing a grammar from text has proven to be a notoriously challenging learning task despite decade...
Stochastic categorial grammars (SCGs) are introduced as a more appropriate formalism for statistical...
Abstract. Although state-of-the-art parsers for natural language are lexicalized, it was recently sh...
Building models of language is a central task in natural language processing. Traditionally, languag...
This paper describes a fully implemented, broad coverage model of human syntactic processing. The mo...
Although state-of-the-art parsers for natural language are lexicalized, it was recently shown that a...
We address the issue of how to associate frequency information with lexicalized grammar formalisms, ...
The task of unsupervised induction of probabilistic context-free grammars (PCFGs) has attracted a lo...
The task of unsupervised induction of probabilistic context-free grammars (PCFGs) has attracted a lo...
The task of unsupervised induction of probabilistic context-free grammars (PCFGs) has attracted a lo...
Language models for speech recognition typically use a probability model of the form Pr(an/a1, a2, ....