Using semi-supervised EM, we learn finegrained but sparse lexical parameters of a generative parsing model (a PCFG) initially estimated over the Penn Treebank. Our lexical parameters employ supertags, which encode complex structural information at the pre-terminal level, and are particularly sparse in labeled data - our goal is to learn these for words that are unseen or rare in the labeled data. In order to guide estimation from unlabeled data, we incorporate both structural and lexical priors from the labeled data. We get a large error reduction in parsing ambiguous structures associated with unseen verbs, the most important case of learning lexico-structural dependencies. We also obtain a statistically significant improvement in labeled ...
There are many methods to improve performance of statistical parsers. Resolving structural ambiguiti...
This paper demonstrates how unsupervised techniques can be used to learn models of deep linguistic s...
Finding the right representations for words is critical for building accurate NLP systems when domai...
This article uses semi-supervised Expectation Maximization (EM) to learn lexico-syntactic dependenci...
Statistical parsers trained on labeled data suffer from sparsity, both grammatical and lexical. For ...
In this dissertation, we have proposed novel methods for robust parsing that integrate the flexibili...
This thesis focuses on robust analysis of natural language semantics. A primary bottleneck for seman...
Treebank parsing can be seen as the search for an optimally refined grammar consistent with a coarse...
Statistical approaches to language learning typically focus on either short-range syntactic dependen...
A probabilistic model of the structural preferences of open-class words is important for accurate pa...
In this work, we apply statistical learning algorithms to Lexicalized Tree Adjoining Grammar (LTAG) ...
After presenting a novel O(n^3) parsing algorithm for dependency grammar, we develop three contrasti...
ynaga at tkl.iis.u-tokyo.ac.jp This paper proposes a method of con-structing an accurate probabilist...
© 2019 Association for Computational Linguistics State-of-the-art LSTM language models trained on la...
We first show how a structural locality bias can improve the accuracy of state-of-the-art dependency...
There are many methods to improve performance of statistical parsers. Resolving structural ambiguiti...
This paper demonstrates how unsupervised techniques can be used to learn models of deep linguistic s...
Finding the right representations for words is critical for building accurate NLP systems when domai...
This article uses semi-supervised Expectation Maximization (EM) to learn lexico-syntactic dependenci...
Statistical parsers trained on labeled data suffer from sparsity, both grammatical and lexical. For ...
In this dissertation, we have proposed novel methods for robust parsing that integrate the flexibili...
This thesis focuses on robust analysis of natural language semantics. A primary bottleneck for seman...
Treebank parsing can be seen as the search for an optimally refined grammar consistent with a coarse...
Statistical approaches to language learning typically focus on either short-range syntactic dependen...
A probabilistic model of the structural preferences of open-class words is important for accurate pa...
In this work, we apply statistical learning algorithms to Lexicalized Tree Adjoining Grammar (LTAG) ...
After presenting a novel O(n^3) parsing algorithm for dependency grammar, we develop three contrasti...
ynaga at tkl.iis.u-tokyo.ac.jp This paper proposes a method of con-structing an accurate probabilist...
© 2019 Association for Computational Linguistics State-of-the-art LSTM language models trained on la...
We first show how a structural locality bias can improve the accuracy of state-of-the-art dependency...
There are many methods to improve performance of statistical parsers. Resolving structural ambiguiti...
This paper demonstrates how unsupervised techniques can be used to learn models of deep linguistic s...
Finding the right representations for words is critical for building accurate NLP systems when domai...