Dropout, a stochastic regularisation technique for training of neural networks, has recently been reinterpreted as a specific type of approximate inference algorithm for Bayesian neural networks. The main contribution of the reinterpretation is in providing a theoretical framework useful for analysing and extending the algorithm. We show that the proposed framework suffers from several issues; from undefined or pathological behaviour of the true posterior related to use of improper priors, to an ill-defined variational objective due to singularity of the approximating distribution relative to the true posterior. Our analysis of the improper log uniform prior used in variational Gaussian dropout suggests the pathologies are generally irredee...
Variational inference (VI) or Variational Bayes (VB) is a popular alternative to MCMC, which doesn\u...
This paper introduces the $\textit{variational Rényi bound}$ (VR) that extends traditional variation...
Variational methods have been previously explored as a tractable approximation to Bayesian inference...
Dropout, a stochastic regularisation technique for training of neural networks, has recently been re...
Gaussian multiplicative noise is commonly used as a stochastic regularisation technique in training ...
Variational dropout (VD) is a generalization of Gaussian dropout, which aims at inferring the poster...
To obtain uncertainty estimates with real-world Bayesian deep learning models, practical inference a...
Recent work has attempted to directly approximate the `function-space' or predictive posterior distr...
We investigate a local reparameterizaton technique for greatly reducing the variance of stochastic g...
© 7th International Conference on Learning Representations, ICLR 2019. All Rights Reserved. Bayesian...
Soft dropout, a generalization of standard “hard” dropout, is introduced to regularize the parameter...
We provide a rigorous analysis of training by variational inference (VI) of Bayesian neural networks...
We show that a neural network with arbitrary depth and non-linearities, with dropout applied before ...
Deep learning tools have gained tremendous attention in applied machine learning. However such tools...
Convolutional neural networks (CNNs) work well on large datasets. But labelled data is hard to colle...
Variational inference (VI) or Variational Bayes (VB) is a popular alternative to MCMC, which doesn\u...
This paper introduces the $\textit{variational Rényi bound}$ (VR) that extends traditional variation...
Variational methods have been previously explored as a tractable approximation to Bayesian inference...
Dropout, a stochastic regularisation technique for training of neural networks, has recently been re...
Gaussian multiplicative noise is commonly used as a stochastic regularisation technique in training ...
Variational dropout (VD) is a generalization of Gaussian dropout, which aims at inferring the poster...
To obtain uncertainty estimates with real-world Bayesian deep learning models, practical inference a...
Recent work has attempted to directly approximate the `function-space' or predictive posterior distr...
We investigate a local reparameterizaton technique for greatly reducing the variance of stochastic g...
© 7th International Conference on Learning Representations, ICLR 2019. All Rights Reserved. Bayesian...
Soft dropout, a generalization of standard “hard” dropout, is introduced to regularize the parameter...
We provide a rigorous analysis of training by variational inference (VI) of Bayesian neural networks...
We show that a neural network with arbitrary depth and non-linearities, with dropout applied before ...
Deep learning tools have gained tremendous attention in applied machine learning. However such tools...
Convolutional neural networks (CNNs) work well on large datasets. But labelled data is hard to colle...
Variational inference (VI) or Variational Bayes (VB) is a popular alternative to MCMC, which doesn\u...
This paper introduces the $\textit{variational Rényi bound}$ (VR) that extends traditional variation...
Variational methods have been previously explored as a tractable approximation to Bayesian inference...