Probabilistic inference on a big data scale is becoming increasingly relevant to both the machine learning and statistics communities. Here we introduce the first fully distributed MCMC algorithm based on stochastic gradients. We argue that stochastic gradient MCMC algorithms are particularly suited for distributed inference because individual chains can draw minibatches from their local pool of data for a flexible amount of time before jumping to or syncing with other chains. This greatly reduces communication overhead and allows adaptive load balancing. Our experiments for LDA on Wikipedia and Pubmed show that relative to the state of the art in distributed MCMC we reduce compute time from 27 hours to half an hour in order to reach the sa...
We consider the distributed stochastic gradient descent problem, where a main node distributes gradi...
In distributed training of deep models, the transmission volume of stochastic gradients (SG) imposes...
International audienceFor large scale inverse problems, inference can be tackled with distributed al...
Probabilistic inference on a big data scale is becoming increasingly relevant to both the machine le...
Probabilistic inference on a big data scale is becoming increasingly relevant to both the machine le...
Probabilistic inference on a big data scale is be-coming increasingly relevant to both the machine l...
Despite the powerful advantages of Bayesian inference such as quantifying uncertainty, ac- curate av...
Learning probability distributions on the weights of neural networks has recently proven beneficial ...
Stochastic gradient MCMC methods, such as stochastic gradient Langevin dynamics (SGLD), employ fast ...
Stochastic gradient MCMC methods, such as stochastic gradient Langevin dynamics (SGLD), employ fast ...
This paper makes two contributions to Bayesian machine learning algorithms. Firstly, we propose stoc...
We propose a stochastic gradient Markov chain Monte Carlo (SG-MCMC) algorithm for scalable inference...
The classical framework on distributed inference considers a set of nodes taking measurements and a ...
We consider the distributed stochastic gradient descent problem, where a main node distributes gradi...
We consider distributed optimization over several devices, each sending incremental model updates to...
We consider the distributed stochastic gradient descent problem, where a main node distributes gradi...
In distributed training of deep models, the transmission volume of stochastic gradients (SG) imposes...
International audienceFor large scale inverse problems, inference can be tackled with distributed al...
Probabilistic inference on a big data scale is becoming increasingly relevant to both the machine le...
Probabilistic inference on a big data scale is becoming increasingly relevant to both the machine le...
Probabilistic inference on a big data scale is be-coming increasingly relevant to both the machine l...
Despite the powerful advantages of Bayesian inference such as quantifying uncertainty, ac- curate av...
Learning probability distributions on the weights of neural networks has recently proven beneficial ...
Stochastic gradient MCMC methods, such as stochastic gradient Langevin dynamics (SGLD), employ fast ...
Stochastic gradient MCMC methods, such as stochastic gradient Langevin dynamics (SGLD), employ fast ...
This paper makes two contributions to Bayesian machine learning algorithms. Firstly, we propose stoc...
We propose a stochastic gradient Markov chain Monte Carlo (SG-MCMC) algorithm for scalable inference...
The classical framework on distributed inference considers a set of nodes taking measurements and a ...
We consider the distributed stochastic gradient descent problem, where a main node distributes gradi...
We consider distributed optimization over several devices, each sending incremental model updates to...
We consider the distributed stochastic gradient descent problem, where a main node distributes gradi...
In distributed training of deep models, the transmission volume of stochastic gradients (SG) imposes...
International audienceFor large scale inverse problems, inference can be tackled with distributed al...