We investigate the problem of training probabilistic context-free grammars on the basis of a distribution defined over an infinite set of trees, by minimizing the cross-entropy. This problem can be seen as a generalization of the well-known maximum likelihood estimator on (finite) tree banks. We prove an unexpected theoretical property of grammars that are trained in this way, namely, we show that the derivational entropy of the grammar takes the same value as the cross-entropy between the input distribution and the grammar itself. We show that the result also holds for the widely applied maximum likelihood estimator on tree banks
The problem of identifying a probabilistic context free grammar has two aspects: the first is determ...
The information theoretical concept of the entropy (channel capacity) of context-free languages and ...
Probabilistic grammars acting as information sources are considered and concepts from information th...
We investigate the problem of training probabilistic context-free grammars on the basis of a distrib...
We investigate the problem of training probabilistic context-free grammars on the basis of a distrib...
In this paper, we consider probabilistic context-free grammars, a class of generative devices that h...
In this paper, we consider probabilistic context-free grammars, a class of generative devices that h...
In this paper, we consider probabilistic context-free grammars, a class of generative devices that h...
In this paper, we consider probabilistic context-free grammars, a class of generative devices that h...
We consider the problem of computing the Kullback-Leibler distance, also called the relative entropy...
We consider the problem of computing the Kullback-Leibler distance, also called the relative entropy...
We consider the problem of computing the Kullback-Leibler distance, also called the relative entropy...
Using an entropy argument, it is shown that stochastic context-free grammars (SCFG's) can model sour...
We consider several empirical estimators for probabilistic context-free grammars, and show that the ...
We consider several empirical estimators for probabilistic context-free grammars, and show that the ...
The problem of identifying a probabilistic context free grammar has two aspects: the first is determ...
The information theoretical concept of the entropy (channel capacity) of context-free languages and ...
Probabilistic grammars acting as information sources are considered and concepts from information th...
We investigate the problem of training probabilistic context-free grammars on the basis of a distrib...
We investigate the problem of training probabilistic context-free grammars on the basis of a distrib...
In this paper, we consider probabilistic context-free grammars, a class of generative devices that h...
In this paper, we consider probabilistic context-free grammars, a class of generative devices that h...
In this paper, we consider probabilistic context-free grammars, a class of generative devices that h...
In this paper, we consider probabilistic context-free grammars, a class of generative devices that h...
We consider the problem of computing the Kullback-Leibler distance, also called the relative entropy...
We consider the problem of computing the Kullback-Leibler distance, also called the relative entropy...
We consider the problem of computing the Kullback-Leibler distance, also called the relative entropy...
Using an entropy argument, it is shown that stochastic context-free grammars (SCFG's) can model sour...
We consider several empirical estimators for probabilistic context-free grammars, and show that the ...
We consider several empirical estimators for probabilistic context-free grammars, and show that the ...
The problem of identifying a probabilistic context free grammar has two aspects: the first is determ...
The information theoretical concept of the entropy (channel capacity) of context-free languages and ...
Probabilistic grammars acting as information sources are considered and concepts from information th...