We formulate natural gradient variational inference (VI), expectation propagation (EP), and posterior linearisation (PL) as extensions of Newton's method for optimising the parameters of a Bayesian posterior distribution. This viewpoint explicitly casts inference algorithms under the framework of numerical optimisation. We show that common approximations to Newton's method from the optimisation literature, namely Gauss-Newton and quasi-Newton methods (e.g., the BFGS algorithm), are still valid under this 'Bayes-Newton' framework. This leads to a suite of novel algorithms which are guaranteed to result in positive semi-definite (PSD) covariance matrices, unlike standard VI and EP. Our unifying viewpoint provides new insights into the connect...
<p>One of the core problems of modern statistics is to approximate difficult-to-compute probability ...
Gaussian processes are distributions over functions that are versatile and mathematically convenient...
Approximating probability densities is a core problem in Bayesian statistics, where the inference in...
We formulate natural gradient variational inference (VI), expectation propagation (EP), and posterio...
Contains fulltext : 83218.pdf (publisher's version ) (Open Access)The results in t...
Variational Bayes methods approximate the posterior density by a family of tractable distributions a...
The Bayesian Conjugate Gradient method (BayesCG) is a probabilistic generalization of the Conjugate ...
We present a framework for approximate Bayesian inference when only a limited number of noisy log-li...
Variational Bayesian inference is an important machine-learning tool that finds application from sta...
Bayesian machine learning has gained tremendous attention in the machine learning community over the...
We advocate an optimization-centric view of Bayesian inference. Our inspiration is the representatio...
Variational Inference (VI) has become a popular technique to approximate difficult-to-compute poster...
Gaussian processes scale prohibitively with the size of the dataset. In response, many approximation...
We develop an optimization algorithm suitable for Bayesian learning in complex models. Our approach ...
This paper presents a scalable approximate Bayesian method for image restoration using total variati...
<p>One of the core problems of modern statistics is to approximate difficult-to-compute probability ...
Gaussian processes are distributions over functions that are versatile and mathematically convenient...
Approximating probability densities is a core problem in Bayesian statistics, where the inference in...
We formulate natural gradient variational inference (VI), expectation propagation (EP), and posterio...
Contains fulltext : 83218.pdf (publisher's version ) (Open Access)The results in t...
Variational Bayes methods approximate the posterior density by a family of tractable distributions a...
The Bayesian Conjugate Gradient method (BayesCG) is a probabilistic generalization of the Conjugate ...
We present a framework for approximate Bayesian inference when only a limited number of noisy log-li...
Variational Bayesian inference is an important machine-learning tool that finds application from sta...
Bayesian machine learning has gained tremendous attention in the machine learning community over the...
We advocate an optimization-centric view of Bayesian inference. Our inspiration is the representatio...
Variational Inference (VI) has become a popular technique to approximate difficult-to-compute poster...
Gaussian processes scale prohibitively with the size of the dataset. In response, many approximation...
We develop an optimization algorithm suitable for Bayesian learning in complex models. Our approach ...
This paper presents a scalable approximate Bayesian method for image restoration using total variati...
<p>One of the core problems of modern statistics is to approximate difficult-to-compute probability ...
Gaussian processes are distributions over functions that are versatile and mathematically convenient...
Approximating probability densities is a core problem in Bayesian statistics, where the inference in...