Discriminative probabilistic models are very popular in NLP because of the latitude they afford in designing features. But training involves complex trade-offs among weights, which can be dangerous: a few highly-indicative features can swamp the contribution of many individually weaker features, causing their weights to be undertrained. Such a model is less robust, for the highly-indicative features may be noisy or missing in the test data. To ameliorate this weight undertraining, we intro-duce several new feature bagging methods, in which separate models are trained on subsets of the original features, and combined using a mixture model or a product of experts. These methods include the logarithmic opinion pools used by Smith et al. (2005)...
Robust Natural Language Processing systems must be able to handle words that are not in their lexico...
Abstract. The major hypothesis that we will be prove in this paper is that unsupervised learning tec...
Since their recent introduction, conditional random fields (CRFs) have been successfully applied to ...
Discriminatively-trained probabilistic models are widely useful because of the latitude they afford ...
In this thesis, we investigate the use of parametric probabilistic models for classification tasks i...
In this thesis, we investigate the use of parametric probabilistic models for classification tasks i...
Breiman's bagging and Freund and Schapire's boosting are recent methods for improving the...
Bagging has been found to be successful in increasing the predictive performance of unstable classif...
We investigate machine learning techniques for coping with highly skewed class distributions in two ...
We investigate machine learning techniques for coping with highly skewed class distributions in two ...
This paper analyses the relation between the use of similarity in MemoryBased Learning and the notio...
Natural language inference (NLI) models are susceptible to learning shortcuts, i.e. decision rules t...
We propose an innovative, effective, and data-agnostic method to train a deep-neural network model w...
In this paper, we propose lazy bagging (LB), which builds bootstrap replicate bags based on the char...
We consider the problem of learning density mixture models for Classification. Traditional learning ...
Robust Natural Language Processing systems must be able to handle words that are not in their lexico...
Abstract. The major hypothesis that we will be prove in this paper is that unsupervised learning tec...
Since their recent introduction, conditional random fields (CRFs) have been successfully applied to ...
Discriminatively-trained probabilistic models are widely useful because of the latitude they afford ...
In this thesis, we investigate the use of parametric probabilistic models for classification tasks i...
In this thesis, we investigate the use of parametric probabilistic models for classification tasks i...
Breiman's bagging and Freund and Schapire's boosting are recent methods for improving the...
Bagging has been found to be successful in increasing the predictive performance of unstable classif...
We investigate machine learning techniques for coping with highly skewed class distributions in two ...
We investigate machine learning techniques for coping with highly skewed class distributions in two ...
This paper analyses the relation between the use of similarity in MemoryBased Learning and the notio...
Natural language inference (NLI) models are susceptible to learning shortcuts, i.e. decision rules t...
We propose an innovative, effective, and data-agnostic method to train a deep-neural network model w...
In this paper, we propose lazy bagging (LB), which builds bootstrap replicate bags based on the char...
We consider the problem of learning density mixture models for Classification. Traditional learning ...
Robust Natural Language Processing systems must be able to handle words that are not in their lexico...
Abstract. The major hypothesis that we will be prove in this paper is that unsupervised learning tec...
Since their recent introduction, conditional random fields (CRFs) have been successfully applied to ...