Variational inference with a factorized Gaussian posterior estimate is a widely-used approach for learning parameters and hidden variables. Empirically, a regularizing effect can be observed that is poorly understood. In this work, we show how mean field inference improves generalization by limiting mutual information between learned parameters and the data through noise. We quantify a maximum capacity when the posterior variance is either fixed or learned and connect it to generalization error, even when the KL-divergence in the objective is scaled by a constant. Our experiments suggest that bounding information between parameters and data effectively regularizes neural networks on both supervised and unsupervised tasks
Gaussian multiplicative noise is commonly used as a stochastic regularisation technique in training ...
Dropout, a stochastic regularisation technique for training of neural networks, has recently been re...
Neural networks can be regarded as statistical models, and can be analysed in a Bayesian framework. ...
Variational inference with a factorized Gaussian posterior estimate is a widely-used approach for le...
Neural networks can be regarded as statistical models, and can be analysed in a Bayesian framework. ...
Abstract Stochastic variational inference makes it possible to approximate posterior distributions i...
We reinterpret multiplicative noise in neural networks as auxiliary random variables that augment th...
Stochastic variational inference makes it possible to approximate posterior distributions induced by...
NOTE: Text or symbols not renderable in plain ASCII are indicated by [...]. Abstract is included in ...
This paper was accepted for publication to Machine Learning (Springer). Overfitting data is a well-k...
International audienceStudies on generalization performance of machine learning algorithms under the...
We propose a simple method that combines neural networks and Gaussian processes. The proposed method...
How can we perform efficient inference and learning in directed probabilistic models, in the presenc...
24 pages, including 2 pages of references and 10 pages of appendixIn machine learning, it is common ...
Dropout, a stochastic regularisation technique for training of neural networks, has recently been re...
Gaussian multiplicative noise is commonly used as a stochastic regularisation technique in training ...
Dropout, a stochastic regularisation technique for training of neural networks, has recently been re...
Neural networks can be regarded as statistical models, and can be analysed in a Bayesian framework. ...
Variational inference with a factorized Gaussian posterior estimate is a widely-used approach for le...
Neural networks can be regarded as statistical models, and can be analysed in a Bayesian framework. ...
Abstract Stochastic variational inference makes it possible to approximate posterior distributions i...
We reinterpret multiplicative noise in neural networks as auxiliary random variables that augment th...
Stochastic variational inference makes it possible to approximate posterior distributions induced by...
NOTE: Text or symbols not renderable in plain ASCII are indicated by [...]. Abstract is included in ...
This paper was accepted for publication to Machine Learning (Springer). Overfitting data is a well-k...
International audienceStudies on generalization performance of machine learning algorithms under the...
We propose a simple method that combines neural networks and Gaussian processes. The proposed method...
How can we perform efficient inference and learning in directed probabilistic models, in the presenc...
24 pages, including 2 pages of references and 10 pages of appendixIn machine learning, it is common ...
Dropout, a stochastic regularisation technique for training of neural networks, has recently been re...
Gaussian multiplicative noise is commonly used as a stochastic regularisation technique in training ...
Dropout, a stochastic regularisation technique for training of neural networks, has recently been re...
Neural networks can be regarded as statistical models, and can be analysed in a Bayesian framework. ...