Abstract—The normalized maximized likelihood (NML) pro-vides the minimax regret solution in universal data compression, gambling, and prediction, and it plays an essential role in the minimum description length (MDL) method of statistical modeling and estimation. Here we show that when the sample space is finite, a generic condition on the linear independence of the component models implies that the normalized maximum likelihood has an exact Bayes-like representation as a mixture of the component models, even in finite samples, though the weights of linear combination may be both positive and negative. This addresses in part the relationship between MDL and Bayes modeling. The representation also has the practical advantage of speeding the ...
The normalized maximum likelihood code length has been widely used in model selection, and its favor...
Learning and compression are driven by the common aim of identifying and exploiting statistical regu...
International audienceThe problem is that of sequential probability forecasting for finite-valued ti...
Abstract—The normalized maximized likelihood (NML) pro-vides the minimax regret solution in universa...
Under the principle of minimum description length, the optimal predictive model maximizes the normal...
: Statistics based inference methods like minimum message length (MML) and minimum description lengt...
Minimum Description Length (MDL) inference is based on the intuition that understanding the availabl...
cCorresponding Author The Minimum Description Length (MDL) principle is an information theoretic app...
Abstract—We analyze the relationship between a Minimum Description Length (MDL) estimator (posterior...
The normalized maximum likelihood distribution achieves minimax coding (log-loss) re-gret given a fi...
Tasks such as data compression and prediction commonly require choosing a probability distribution o...
An applied problem is discussed in which two nested psychological models of retention are compared u...
We study online learning under logarithmic loss with regular parametric models. In this setting, eac...
Due to non-regularity of the finite mixture of normal dis-tributions in both mean and variance, the ...
Abstract: This paper deals with a computational aspect of the Bayesian analysis of statisti-cal mode...
The normalized maximum likelihood code length has been widely used in model selection, and its favor...
Learning and compression are driven by the common aim of identifying and exploiting statistical regu...
International audienceThe problem is that of sequential probability forecasting for finite-valued ti...
Abstract—The normalized maximized likelihood (NML) pro-vides the minimax regret solution in universa...
Under the principle of minimum description length, the optimal predictive model maximizes the normal...
: Statistics based inference methods like minimum message length (MML) and minimum description lengt...
Minimum Description Length (MDL) inference is based on the intuition that understanding the availabl...
cCorresponding Author The Minimum Description Length (MDL) principle is an information theoretic app...
Abstract—We analyze the relationship between a Minimum Description Length (MDL) estimator (posterior...
The normalized maximum likelihood distribution achieves minimax coding (log-loss) re-gret given a fi...
Tasks such as data compression and prediction commonly require choosing a probability distribution o...
An applied problem is discussed in which two nested psychological models of retention are compared u...
We study online learning under logarithmic loss with regular parametric models. In this setting, eac...
Due to non-regularity of the finite mixture of normal dis-tributions in both mean and variance, the ...
Abstract: This paper deals with a computational aspect of the Bayesian analysis of statisti-cal mode...
The normalized maximum likelihood code length has been widely used in model selection, and its favor...
Learning and compression are driven by the common aim of identifying and exploiting statistical regu...
International audienceThe problem is that of sequential probability forecasting for finite-valued ti...