We present a new family of information-theoretic generalization bounds, in which the training loss and the population loss are compared through a jointly convex function. This function is upper-bounded in terms of the disintegrated, samplewise, evaluated conditional mutual information (CMI), an information measure that depends on the losses incurred by the selected hypothesis, rather than on the hypothesis itself, as is common in probably approximately correct (PAC)-Bayesian results. We demonstrate the generality of this framework by recovering and extending previously known information-theoretic bounds. Furthermore, using the evaluated CMI, we derive a samplewise, average version of Seeger\u27s PAC-Bayesian bound, where the convex function...
PAC-Bayesian bounds are known to be tight and informative when studying the generalization ability o...
PAC-Bayesian bounds are known to be tight and informative when studying the generalization ability o...
Neural network learning rules can be viewed as statistical estimators. They should be studied in Bay...
Machine learning has achieved impressive feats in numerous domains, largely driven by the emergence ...
During the past decade, machine learning techniques have achieved impressive results in a number of ...
We present a general approach to deriving bounds on the generalization error of randomized learning ...
29 pagesWe derive generic information-theoretic and PAC-Bayesian generalization bounds involving an ...
Some of the tightest information-theoretic generalization bounds depend on the average information b...
We derive information theoretic generalization bounds for supervised learning algorithms based on a ...
Generalization error bounds are critical to understanding the performance of machine learning models...
We give a novel, unified derivation of conditional PAC-Bayesian and mutual information (MI) generali...
International audiencePAC-Bayesian bounds are known to be tight and informative when studying the ge...
International audiencePAC-Bayesian bounds are known to be tight and informative when studying the ge...
International audiencePAC-Bayesian bounds are known to be tight and informative when studying the ge...
International audiencePAC-Bayesian bounds are known to be tight and informative when studying the ge...
PAC-Bayesian bounds are known to be tight and informative when studying the generalization ability o...
PAC-Bayesian bounds are known to be tight and informative when studying the generalization ability o...
Neural network learning rules can be viewed as statistical estimators. They should be studied in Bay...
Machine learning has achieved impressive feats in numerous domains, largely driven by the emergence ...
During the past decade, machine learning techniques have achieved impressive results in a number of ...
We present a general approach to deriving bounds on the generalization error of randomized learning ...
29 pagesWe derive generic information-theoretic and PAC-Bayesian generalization bounds involving an ...
Some of the tightest information-theoretic generalization bounds depend on the average information b...
We derive information theoretic generalization bounds for supervised learning algorithms based on a ...
Generalization error bounds are critical to understanding the performance of machine learning models...
We give a novel, unified derivation of conditional PAC-Bayesian and mutual information (MI) generali...
International audiencePAC-Bayesian bounds are known to be tight and informative when studying the ge...
International audiencePAC-Bayesian bounds are known to be tight and informative when studying the ge...
International audiencePAC-Bayesian bounds are known to be tight and informative when studying the ge...
International audiencePAC-Bayesian bounds are known to be tight and informative when studying the ge...
PAC-Bayesian bounds are known to be tight and informative when studying the generalization ability o...
PAC-Bayesian bounds are known to be tight and informative when studying the generalization ability o...
Neural network learning rules can be viewed as statistical estimators. They should be studied in Bay...