We present paired learning and inference algorithms for significantly reducing computation and increasing speed of the vector dot products in the classifiers that are at the heart of many NLP components. This is accomplished by partition-ing the features into a sequence of templates which are ordered such that high confidence can often be reached using only a small fraction of all features. Pa-rameter estimation is arranged to maximize accuracy and early confidence in this sequence. We present experiments in left-to-right part-of-speech tagging on WSJ, demonstrating that we can preserve accuracy above 97 % with over a five-fold re-duction in run-time.
The classification learning task requires selection of a subset of features to represent patterns to...
Discriminative probabilistic models are very popular in NLP because of the latitude they afford in d...
Learning a sequence classifier means learning to predict a sequence of output tags based on a set of...
We present paired learning and inference algorithms for significantly reducing computation and incre...
This paper proposes a method that speeds up a classifier trained with many con-junctive features: co...
Sequence Tagging, including part of speech tagging, chunking and named entity recognition, is an imp...
Some machine learning tasks have a complex output, rather than a real number or a class. Those outpu...
Exponentiated Gradient (EG) updates were originally introduced in (Kivinen and Warmuth, 1997) in the...
Recently, significant progress has been made on learning structured predictors via coordinated train...
Symbolic machine-learning classifiers are known to suffer from near-sightedness when performing sequ...
Abstract. In most real-world information processing problems, data is not a free resource; its acqui...
Discriminative learning framework is one of the very successful fields of machine learn-ing. The met...
Training speed and accuracy are two major concerns of large-scale natural language processing system...
Training speed and accuracy are two major concerns of large-scale natural language processing system...
Training speed and accuracy are two major concerns of large-scale natural language processing system...
The classification learning task requires selection of a subset of features to represent patterns to...
Discriminative probabilistic models are very popular in NLP because of the latitude they afford in d...
Learning a sequence classifier means learning to predict a sequence of output tags based on a set of...
We present paired learning and inference algorithms for significantly reducing computation and incre...
This paper proposes a method that speeds up a classifier trained with many con-junctive features: co...
Sequence Tagging, including part of speech tagging, chunking and named entity recognition, is an imp...
Some machine learning tasks have a complex output, rather than a real number or a class. Those outpu...
Exponentiated Gradient (EG) updates were originally introduced in (Kivinen and Warmuth, 1997) in the...
Recently, significant progress has been made on learning structured predictors via coordinated train...
Symbolic machine-learning classifiers are known to suffer from near-sightedness when performing sequ...
Abstract. In most real-world information processing problems, data is not a free resource; its acqui...
Discriminative learning framework is one of the very successful fields of machine learn-ing. The met...
Training speed and accuracy are two major concerns of large-scale natural language processing system...
Training speed and accuracy are two major concerns of large-scale natural language processing system...
Training speed and accuracy are two major concerns of large-scale natural language processing system...
The classification learning task requires selection of a subset of features to represent patterns to...
Discriminative probabilistic models are very popular in NLP because of the latitude they afford in d...
Learning a sequence classifier means learning to predict a sequence of output tags based on a set of...