Combining a naive Bayes classifier with the EM algorithm is one of the promising ap-proaches for making use of unlabeled data for disambiguation tasks when using local con-text features including word sense disambigua-tion and spelling correction. However, the use of unlabeled data via the basic EM algorithm often causes disastrous performance degrada-tion instead of improving classification perfor-mance, resulting in poor classification perfor-mance on average. In this study, we introduce a class distribution constraint into the iteration process of the EM algorithm. This constraint keeps the class distribution of unlabeled data consistent with the class distribution estimated from labeled data, preventing the EM algorithm from converging ...
Naive Bayes is often used as a baseline in text classification because it is fast and easy to implem...
In applications of data mining characterized by highly skewed misclassification costs certain types ...
Abstract. This paper studies the performance of various classifiers for Word Sense Disambiguation co...
This paper shows that the accuracy of learned text classifiers can be improved by augmenting a small...
In many important text classification problems, acquiring class labels for training documents is cos...
We introduce a Bayesian model, BayesANIL, that is capable of estimating uncertainties associated wit...
Abstract. This paper presents the method of significantly improving conventional Bayesian statistica...
Multinomial Naive Bayes with Expectation Maximization (MNB-EM) is a standard semi-supervised learnin...
It is an actual and challenging issue to learn cost-sensitive models from those datasets that are wi...
International audienceThis paper investigates a new approach for training discriminant classifiers w...
We introduce a new method for improving poor perfor-mance of classi£ers due to a small training set....
This report considers a paper [1] that com-bines the Expectation-Maximisation (EM) algo-rithm and a ...
document are those of the author and should not be interpreted as representing the official policies...
One of the advantages of supervised learning is that the final error metric is available during trai...
We propose a novel dictionary-based learning method for ambiguously labeled multiclass classificatio...
Naive Bayes is often used as a baseline in text classification because it is fast and easy to implem...
In applications of data mining characterized by highly skewed misclassification costs certain types ...
Abstract. This paper studies the performance of various classifiers for Word Sense Disambiguation co...
This paper shows that the accuracy of learned text classifiers can be improved by augmenting a small...
In many important text classification problems, acquiring class labels for training documents is cos...
We introduce a Bayesian model, BayesANIL, that is capable of estimating uncertainties associated wit...
Abstract. This paper presents the method of significantly improving conventional Bayesian statistica...
Multinomial Naive Bayes with Expectation Maximization (MNB-EM) is a standard semi-supervised learnin...
It is an actual and challenging issue to learn cost-sensitive models from those datasets that are wi...
International audienceThis paper investigates a new approach for training discriminant classifiers w...
We introduce a new method for improving poor perfor-mance of classi£ers due to a small training set....
This report considers a paper [1] that com-bines the Expectation-Maximisation (EM) algo-rithm and a ...
document are those of the author and should not be interpreted as representing the official policies...
One of the advantages of supervised learning is that the final error metric is available during trai...
We propose a novel dictionary-based learning method for ambiguously labeled multiclass classificatio...
Naive Bayes is often used as a baseline in text classification because it is fast and easy to implem...
In applications of data mining characterized by highly skewed misclassification costs certain types ...
Abstract. This paper studies the performance of various classifiers for Word Sense Disambiguation co...