In this paper we analyze the average behavior of the Bayes-optimal and Gibbs learning algorithms. We do this both for off-training-set error and conventional IID error (for which test sets overlap with training sets). For the IID case we provide a major extension to one of the better known results of [7]. We also show that expected IID test set error is a non-increasing function of training set size for either algorithm. On the other hand, as we show, the expected off training-set error for both learning algorithms can increase with training set size, for non-uniform sampling distributions. We characterize what relationship the sampling distribution must have with the prior for such an increase. We show in particular that for uniform sampli...
In order to compare learning algorithms, experimental results reported in the machine learning liter...
We characterize the convergence of the Gibbs sampler which samples from the joint posterior distribu...
We have a probabilistic statistical model which is required to adapt in the light of observed cases...
In this paper we analyze the average behavior of the Bayes-optimal and Gibbs learning algorithms. We...
In this paper we analyze the average off-training-set behavior of the Bayes-optimal and Gibbs learni...
Various approaches have been developed to upper bound the generalization error of a supervised learn...
We investigate why discretization is effective in naive-Bayes learning. We prove a theorem that iden...
Bayesian predictions are stochastic just like predictions of any other inference scheme that general...
Various approaches have been developed to upper bound the generalization error of a supervised lear...
A general approach to Bayesian learning revisits some classical results, which study which functiona...
Abstract Bayesian predictions are stochastic just like predictions of any other inference scheme tha...
Many classification problems are solved by aggregating the output of a group of distinct predictors....
We study the interaction between input distributions, learning algo-rithms, and finite sample sizes ...
this article we investigate the relationship between the two popular algorithms, the EM algorithm an...
We study the interaction between input distributions, learning algorithms and finite sample sizes in...
In order to compare learning algorithms, experimental results reported in the machine learning liter...
We characterize the convergence of the Gibbs sampler which samples from the joint posterior distribu...
We have a probabilistic statistical model which is required to adapt in the light of observed cases...
In this paper we analyze the average behavior of the Bayes-optimal and Gibbs learning algorithms. We...
In this paper we analyze the average off-training-set behavior of the Bayes-optimal and Gibbs learni...
Various approaches have been developed to upper bound the generalization error of a supervised learn...
We investigate why discretization is effective in naive-Bayes learning. We prove a theorem that iden...
Bayesian predictions are stochastic just like predictions of any other inference scheme that general...
Various approaches have been developed to upper bound the generalization error of a supervised lear...
A general approach to Bayesian learning revisits some classical results, which study which functiona...
Abstract Bayesian predictions are stochastic just like predictions of any other inference scheme tha...
Many classification problems are solved by aggregating the output of a group of distinct predictors....
We study the interaction between input distributions, learning algo-rithms, and finite sample sizes ...
this article we investigate the relationship between the two popular algorithms, the EM algorithm an...
We study the interaction between input distributions, learning algorithms and finite sample sizes in...
In order to compare learning algorithms, experimental results reported in the machine learning liter...
We characterize the convergence of the Gibbs sampler which samples from the joint posterior distribu...
We have a probabilistic statistical model which is required to adapt in the light of observed cases...