Recent work has attempted to directly approximate the `function-space' or predictive posterior distribution of Bayesian models, without approximating the posterior distribution over the parameters. This is appealing in e.g. Bayesian neural networks, where we only need the former, and the latter is hard to represent. In this work, we highlight some advantages and limitations of employing the Kullback-Leibler divergence in this setting. For example, we show that minimizing the KL divergence between a wide class of parametric distributions and the posterior induced by a (non-degenerate) Gaussian process prior leads to an ill-defined objective function. Then, we propose (featurized) Bayesian linear regression as a benchmark for `function-space'...
Kullback-Leibler (KL) divergence is widely used for variational inference of Bayesian Neural Network...
Variational inference is an optimization-based method for approximating the posterior distribution o...
The results in this thesis are based on applications of the expectation propagation algorithm to app...
Dropout, a stochastic regularisation technique for training of neural networks, has recently been re...
Dropout, a stochastic regularisation technique for training of neural networks, has recently been re...
Variational Bayesian inference is an important machine-learning tool that finds application from sta...
Having access to accurate confidence levels along with the predictions allows to determine whether m...
Gaussian multiplicative noise is commonly used as a stochastic regularisation technique in training ...
Variational inference is an optimization-based method for approximating the posterior distribution o...
We propose a general algorithm for approximating nonstandard Bayesian posterior distributions. The a...
Variational methods for approximate Bayesian inference provide fast, flexible, deterministic alterna...
<p>One of the core problems of modern statistics is to approximate difficult-to-compute probability ...
We provide a rigorous analysis of training by variational inference (VI) of Bayesian neural networks...
We present a novel method for approximate inference. Using some of the constructs from expectation p...
We advocate an optimization-centric view of Bayesian inference. Our inspiration is the representatio...
Kullback-Leibler (KL) divergence is widely used for variational inference of Bayesian Neural Network...
Variational inference is an optimization-based method for approximating the posterior distribution o...
The results in this thesis are based on applications of the expectation propagation algorithm to app...
Dropout, a stochastic regularisation technique for training of neural networks, has recently been re...
Dropout, a stochastic regularisation technique for training of neural networks, has recently been re...
Variational Bayesian inference is an important machine-learning tool that finds application from sta...
Having access to accurate confidence levels along with the predictions allows to determine whether m...
Gaussian multiplicative noise is commonly used as a stochastic regularisation technique in training ...
Variational inference is an optimization-based method for approximating the posterior distribution o...
We propose a general algorithm for approximating nonstandard Bayesian posterior distributions. The a...
Variational methods for approximate Bayesian inference provide fast, flexible, deterministic alterna...
<p>One of the core problems of modern statistics is to approximate difficult-to-compute probability ...
We provide a rigorous analysis of training by variational inference (VI) of Bayesian neural networks...
We present a novel method for approximate inference. Using some of the constructs from expectation p...
We advocate an optimization-centric view of Bayesian inference. Our inspiration is the representatio...
Kullback-Leibler (KL) divergence is widely used for variational inference of Bayesian Neural Network...
Variational inference is an optimization-based method for approximating the posterior distribution o...
The results in this thesis are based on applications of the expectation propagation algorithm to app...