Maximum Entropy Principle has been used successfully in various NLP tasks. In this paper we propose a forward translation model consisting of a set of maximum entropy classifiers: a separate classifier is trained for each (sufficiently frequent) source-side lemma. In this way the estimates of translation probabilities can be sensitive to a large number of features derived from the source sentence (including non-local features, features making use of sentence syntactic structure, etc.). When integrated into English-to- Czech dependency-based translation scenario implemented in the TectoMT framework, the new translation model significantly outperforms the baseline model (MLE) in terms of BLEU. The performance is further boosted in a configura...
In this paper we compare two approaches to natural language understanding (NLU). The first approach ...
In this paper, we des ribe a unied probabilisti framework for statisti al language modeling|the lat...
The conventional n-gram language model exploits only the immediate context of historical words witho...
Maximum Entropy Principle has been used successfully in various NLP tasks. In this paper we propose ...
This paper proposes a learning method of translation rules from parallel corpora. This method applie...
We present a framework for statistical machine translation of natural languages based on direct maxi...
In this paper we present an unsupervised method for learning a model to distinguish between ambiguou...
We propose a novel reordering model for phrase-based statistical machine transla-tion (SMT) that use...
Maximum entropy approaches for sequences tagging and conditional random fields in particular have sh...
This paper proposes a feature extraction algorithm based on the maximum entropy phrase reordering mo...
Many problems in natural language processing can be viewed as linguistic classification problems, in...
This thesis demonstrates that several important kinds of natural language ambiguities can be resolve...
This paper proposes a novel maximum en-tropy based rule selection (MERS) model for syntax-based stat...
Recent work has shown that translating seg-mentation lattices (lattices that encode alterna-tive way...
This thesis demonstrates that several important kinds of natural language ambiguities can be resolve...
In this paper we compare two approaches to natural language understanding (NLU). The first approach ...
In this paper, we des ribe a unied probabilisti framework for statisti al language modeling|the lat...
The conventional n-gram language model exploits only the immediate context of historical words witho...
Maximum Entropy Principle has been used successfully in various NLP tasks. In this paper we propose ...
This paper proposes a learning method of translation rules from parallel corpora. This method applie...
We present a framework for statistical machine translation of natural languages based on direct maxi...
In this paper we present an unsupervised method for learning a model to distinguish between ambiguou...
We propose a novel reordering model for phrase-based statistical machine transla-tion (SMT) that use...
Maximum entropy approaches for sequences tagging and conditional random fields in particular have sh...
This paper proposes a feature extraction algorithm based on the maximum entropy phrase reordering mo...
Many problems in natural language processing can be viewed as linguistic classification problems, in...
This thesis demonstrates that several important kinds of natural language ambiguities can be resolve...
This paper proposes a novel maximum en-tropy based rule selection (MERS) model for syntax-based stat...
Recent work has shown that translating seg-mentation lattices (lattices that encode alterna-tive way...
This thesis demonstrates that several important kinds of natural language ambiguities can be resolve...
In this paper we compare two approaches to natural language understanding (NLU). The first approach ...
In this paper, we des ribe a unied probabilisti framework for statisti al language modeling|the lat...
The conventional n-gram language model exploits only the immediate context of historical words witho...