Recent studies have empirically investigated different methods to train stochastic neural networks on a classification task by optimising a PAC-Bayesian bound via stochastic gradient descent. Most of these procedures need to replace the misclassification error with a surrogate loss, leading to a mismatch between the optimisation objective and the actual generalisation bound. The present paper proposes a novel training algorithm that optimises the PAC-Bayesian bound, without relying on any surrogate loss. Empirical results show that this approach outperforms currently available PAC-Bayesian training methods
Conventional training methods for neural networks involve starting al a random location in the solut...
We focus on a specific class of shallow neural networks with a single hidden layer, namely those wit...
In this paper we further develop the idea that the PAC-Bayes prior can be defined based on the data-...
Recent studies have empirically investigated different methods to train stochastic neural networks o...
The limit of infinite width allows for substantial simplifications in the analytical study of overpa...
This paper presents an empirical study regarding training probabilistic neural networks using traini...
We make three related contributions motivated by the challenge of training stochastic neural network...
We establish a disintegrated PAC-Bayesian bound, for classifiers that are trained via continuous-tim...
PAC-Bayesian bounds are known to be tight and informative when studying the generalization ability o...
We present a new approach to bounding the true error rate of a continuous valued classifier based up...
International audienceA learning method is self-certified if it uses all available data to simultane...
International audiencePAC-Bayesian bounds are known to be tight and informative when studying the ge...
Institute for Adaptive and Neural ComputationNon-parametric models and techniques enjoy a growing po...
Recent works have investigated deep learning models trained by optimising PAC-Bayes bounds, with pri...
This paper introduces a new neural network based prior for real valued functions on $\mathbb R^d$ wh...
Conventional training methods for neural networks involve starting al a random location in the solut...
We focus on a specific class of shallow neural networks with a single hidden layer, namely those wit...
In this paper we further develop the idea that the PAC-Bayes prior can be defined based on the data-...
Recent studies have empirically investigated different methods to train stochastic neural networks o...
The limit of infinite width allows for substantial simplifications in the analytical study of overpa...
This paper presents an empirical study regarding training probabilistic neural networks using traini...
We make three related contributions motivated by the challenge of training stochastic neural network...
We establish a disintegrated PAC-Bayesian bound, for classifiers that are trained via continuous-tim...
PAC-Bayesian bounds are known to be tight and informative when studying the generalization ability o...
We present a new approach to bounding the true error rate of a continuous valued classifier based up...
International audienceA learning method is self-certified if it uses all available data to simultane...
International audiencePAC-Bayesian bounds are known to be tight and informative when studying the ge...
Institute for Adaptive and Neural ComputationNon-parametric models and techniques enjoy a growing po...
Recent works have investigated deep learning models trained by optimising PAC-Bayes bounds, with pri...
This paper introduces a new neural network based prior for real valued functions on $\mathbb R^d$ wh...
Conventional training methods for neural networks involve starting al a random location in the solut...
We focus on a specific class of shallow neural networks with a single hidden layer, namely those wit...
In this paper we further develop the idea that the PAC-Bayes prior can be defined based on the data-...