During the past decade, machine learning techniques have achieved impressive results in a number of domains. Many of the success stories have made use of deep neural networks, a class of functions that boasts high complexity. Classical results that mathematically guarantee that a learning algorithm generalizes, i.e., performs as well on unseen data as on training data, typically rely on bounding the complexity and expressiveness of the functions that are used. As a consequence of this, they yield overly pessimistic results when applied to modern machine learning algorithms, and fail to explain why they generalize.This discrepancy between theoretical explanations and practical success has spurred a flurry of research activity into new genera...
In the context of assessing the generalization abilities of a randomized model or learning algorithm...
This thesis presents a new theory of generalization in neural network types of learning machines. Th...
This is the final version. Available from ICLR via the link in this recordDeep neural networks (DNNs...
Machine learning has achieved impressive feats in numerous domains, largely driven by the emergence ...
Some of the tightest information-theoretic generalization bounds depend on the average information b...
Generalization error bounds are critical to understanding the performance of machine learning models...
We present a general approach to deriving bounds on the generalization error of randomized learning ...
In this work, we construct generalization bounds to understand existing learning algorithms and prop...
We present a new family of information-theoretic generalization bounds, in which the training loss a...
Deep learning has transformed computer vision, natural language processing, and speech recognition. ...
With a direct analysis of neural networks, this paper presents a mathematically tight generalization...
While there has been progress in developing non-vacuous generalization bounds for deep neural networ...
Existing generalization bounds fail to explain crucial factors that drive generalization of modern n...
In the context of assessing the generalization abilities of a randomized model or learning algorithm...
We consider information-theoretic bounds on the expected generalization error for statistical learni...
In the context of assessing the generalization abilities of a randomized model or learning algorithm...
This thesis presents a new theory of generalization in neural network types of learning machines. Th...
This is the final version. Available from ICLR via the link in this recordDeep neural networks (DNNs...
Machine learning has achieved impressive feats in numerous domains, largely driven by the emergence ...
Some of the tightest information-theoretic generalization bounds depend on the average information b...
Generalization error bounds are critical to understanding the performance of machine learning models...
We present a general approach to deriving bounds on the generalization error of randomized learning ...
In this work, we construct generalization bounds to understand existing learning algorithms and prop...
We present a new family of information-theoretic generalization bounds, in which the training loss a...
Deep learning has transformed computer vision, natural language processing, and speech recognition. ...
With a direct analysis of neural networks, this paper presents a mathematically tight generalization...
While there has been progress in developing non-vacuous generalization bounds for deep neural networ...
Existing generalization bounds fail to explain crucial factors that drive generalization of modern n...
In the context of assessing the generalization abilities of a randomized model or learning algorithm...
We consider information-theoretic bounds on the expected generalization error for statistical learni...
In the context of assessing the generalization abilities of a randomized model or learning algorithm...
This thesis presents a new theory of generalization in neural network types of learning machines. Th...
This is the final version. Available from ICLR via the link in this recordDeep neural networks (DNNs...