We study online learning under logarithmic loss with regular parametric models. In this setting, each strategy corresponds to a joint distribution on sequences. The minimax optimal strategy is the normalized maximum likelihood (NML) strategy. We show that the sequential NML (SNML) strategy predicts minimax optimally (i.e., as NML) if and only if the joint distribution on sequences defined by SNML is exchangeable. This property also characterizes the optimality of a Bayesian prediction strategy. In that case, the optimal prior distribution is Jeffreys prior for a broad class of parametric models for which the maximum likelihood estimator is asymptotically normal. The optimal prediction strategy, NML, depends on the number n of rounds of the ...
This paper studies the asymptotic behavior of Bayesian learning processes for general finite-player...
We present a competitive analysis of Bayesian learning algorithms in the online learning setting and...
Various optimality properties of universal sequence predictors based on Bayes-mixtures in general, a...
We study online learning under logarithmic loss with regular parametric models. In this setting, eac...
We study online prediction of individual sequences under logarithmic loss with parametric experts. T...
We study online learning under logarithmic loss with regular parametric models. Hedayati and Bartlet...
The paper considers sequential prediction of individual sequences with log loss (online density esti...
The normalized maximum likelihood model achieves the minimax coding (log-loss) regret for data of fi...
Various optimality properties of universal sequence predictors based on Bayes-mixtures in general, a...
Minimum description length (MDL) is an important principle for induction and prediction, with strong...
The normalized maximum likelihood distribution achieves minimax coding (log-loss) re-gret given a fi...
We introduce an information theoretic criterion for Bayesian network structure learning which we cal...
AbstractWe consider the problem of learning Bayesian network models in a non-informative setting, wh...
We present a competitive analysis of some non-parametric Bayesian algorithms in a worst-case online ...
. We consider the game of sequentially assigning probabilities to future data based on past observat...
This paper studies the asymptotic behavior of Bayesian learning processes for general finite-player...
We present a competitive analysis of Bayesian learning algorithms in the online learning setting and...
Various optimality properties of universal sequence predictors based on Bayes-mixtures in general, a...
We study online learning under logarithmic loss with regular parametric models. In this setting, eac...
We study online prediction of individual sequences under logarithmic loss with parametric experts. T...
We study online learning under logarithmic loss with regular parametric models. Hedayati and Bartlet...
The paper considers sequential prediction of individual sequences with log loss (online density esti...
The normalized maximum likelihood model achieves the minimax coding (log-loss) regret for data of fi...
Various optimality properties of universal sequence predictors based on Bayes-mixtures in general, a...
Minimum description length (MDL) is an important principle for induction and prediction, with strong...
The normalized maximum likelihood distribution achieves minimax coding (log-loss) re-gret given a fi...
We introduce an information theoretic criterion for Bayesian network structure learning which we cal...
AbstractWe consider the problem of learning Bayesian network models in a non-informative setting, wh...
We present a competitive analysis of some non-parametric Bayesian algorithms in a worst-case online ...
. We consider the game of sequentially assigning probabilities to future data based on past observat...
This paper studies the asymptotic behavior of Bayesian learning processes for general finite-player...
We present a competitive analysis of Bayesian learning algorithms in the online learning setting and...
Various optimality properties of universal sequence predictors based on Bayes-mixtures in general, a...