Best Paper AwardInternational audienceOne way to avoid overfitting in machine learning is to use model parameters distributed according to a Bayesian posterior given the data, rather than the maximum likelihood estimator. Stochastic gradient Langevin dynamics (SGLD) is one algorithm to approximate such Bayesian posteriors for large models and datasets. SGLD is a standard stochastic gradient descent to which is added a controlled amount of noise, specifically scaled so that the parameter converges in law to the posterior distribution [WT11, TTV16]. The posterior predictive distribution can be approximated by an ensemble of samples from the trajectory. Choice of the variance of the noise is known to impact the practical behavior of SGLD: for ...
Stochastic-gradient sampling methods are often used to perform Bayesian inference on neural networks...
Stochastic optimization methods (Robbins-Monro algorithms) have been very successful for large scale...
In this thesis, we study model parameterization for deep learning applications. Part of the mathemat...
Best Paper AwardInternational audienceOne way to avoid overfitting in machine learning is to use mod...
Best Paper AwardInternational audienceOne way to avoid overfitting in machine learning is to use mod...
Best Paper AwardInternational audienceOne way to avoid overfitting in machine learning is to use mod...
Effective training of deep neural networks suffers from two main issues. The first is that the param...
International audienceStochastic Gradient Langevin Dynamics (SGLD) has emerged as a key MCMC algorit...
Year after years, the amount of data that we continuously generate is increasing. When this situatio...
Year after years, the amount of data that we continuously generate is increasing. When this situatio...
We propose an adaptively weighted stochastic gradient Langevin dynamics algorithm (SGLD), so-called ...
In this paper we propose a new framework for learning from large scale datasets based on iterative l...
Stochastic gradient MCMC methods, such as stochastic gradient Langevin dynamics (SGLD), employ fast ...
Stochastic gradient MCMC methods, such as stochastic gradient Langevin dynamics (SGLD), employ fast ...
Stochastic-gradient sampling methods are often used to perform Bayesian inference on neural networks...
Stochastic-gradient sampling methods are often used to perform Bayesian inference on neural networks...
Stochastic optimization methods (Robbins-Monro algorithms) have been very successful for large scale...
In this thesis, we study model parameterization for deep learning applications. Part of the mathemat...
Best Paper AwardInternational audienceOne way to avoid overfitting in machine learning is to use mod...
Best Paper AwardInternational audienceOne way to avoid overfitting in machine learning is to use mod...
Best Paper AwardInternational audienceOne way to avoid overfitting in machine learning is to use mod...
Effective training of deep neural networks suffers from two main issues. The first is that the param...
International audienceStochastic Gradient Langevin Dynamics (SGLD) has emerged as a key MCMC algorit...
Year after years, the amount of data that we continuously generate is increasing. When this situatio...
Year after years, the amount of data that we continuously generate is increasing. When this situatio...
We propose an adaptively weighted stochastic gradient Langevin dynamics algorithm (SGLD), so-called ...
In this paper we propose a new framework for learning from large scale datasets based on iterative l...
Stochastic gradient MCMC methods, such as stochastic gradient Langevin dynamics (SGLD), employ fast ...
Stochastic gradient MCMC methods, such as stochastic gradient Langevin dynamics (SGLD), employ fast ...
Stochastic-gradient sampling methods are often used to perform Bayesian inference on neural networks...
Stochastic-gradient sampling methods are often used to perform Bayesian inference on neural networks...
Stochastic optimization methods (Robbins-Monro algorithms) have been very successful for large scale...
In this thesis, we study model parameterization for deep learning applications. Part of the mathemat...