Gradient-descent-based algorithms and their stochastic versions have widespread applications in machine learning and statistical inference. In this work, we carry out an analytic study of the performance of the algorithm most commonly considered in physics, the Langevin algorithm, in the context of noisy high-dimensional inference. We employ the Langevin algorithm to sample the posterior probability measure for the spiked mixed matrix-tensor model. The typical behavior of this algorithm is described by a system of integrodifferential equations that we call the Langevin state evolution, whose solution is compared with the one of the state evolution of approximate message passing (AMP). Our results show that, remarkably, the algorithmic thres...
This thesis focuses on adaptive Stochastic Gradient Langevin Dynamics (SGLD) algorithms to solve opt...
Stochastic gradient MCMC methods, such as stochastic gradient Langevin dynamics (SGLD), employ fast ...
In this paper, we study the computational complexity of sampling from a Bayesian posterior (or pseud...
Gradient-descent-based algorithms and their stochastic versions have widespread applications in mach...
Gradient-descent-based algorithms and their stochastic versions have widespread applications in mach...
International audienceStochastic Gradient Langevin Dynamics (SGLD) has emerged as a key MCMC algorit...
We introduce a novel and efficient algorithm called the stochastic approximate gradient descent (SAG...
International audienceIn this work we analyse quantitatively the interplay between the loss landscap...
Best Paper AwardInternational audienceOne way to avoid overfitting in machine learning is to use mod...
In this paper we propose a new framework for learning from large scale datasets based on iterative l...
Applying standard Markov chain Monte Carlo (MCMC) algorithms to large data sets is computationally e...
Applying standard Markov chain Monte Carlo (MCMC) algorithms to large data sets is computationally e...
This paper proposes a new sampling scheme based on Langevin dynamics that is applicable within pseud...
Applying standard Markov chain Monte Carlo (MCMC) algorithms to large data sets is computationally e...
The Metropolis-adjusted Langevin (MALA) algorithm is a sampling algorithm which makes local moves by...
This thesis focuses on adaptive Stochastic Gradient Langevin Dynamics (SGLD) algorithms to solve opt...
Stochastic gradient MCMC methods, such as stochastic gradient Langevin dynamics (SGLD), employ fast ...
In this paper, we study the computational complexity of sampling from a Bayesian posterior (or pseud...
Gradient-descent-based algorithms and their stochastic versions have widespread applications in mach...
Gradient-descent-based algorithms and their stochastic versions have widespread applications in mach...
International audienceStochastic Gradient Langevin Dynamics (SGLD) has emerged as a key MCMC algorit...
We introduce a novel and efficient algorithm called the stochastic approximate gradient descent (SAG...
International audienceIn this work we analyse quantitatively the interplay between the loss landscap...
Best Paper AwardInternational audienceOne way to avoid overfitting in machine learning is to use mod...
In this paper we propose a new framework for learning from large scale datasets based on iterative l...
Applying standard Markov chain Monte Carlo (MCMC) algorithms to large data sets is computationally e...
Applying standard Markov chain Monte Carlo (MCMC) algorithms to large data sets is computationally e...
This paper proposes a new sampling scheme based on Langevin dynamics that is applicable within pseud...
Applying standard Markov chain Monte Carlo (MCMC) algorithms to large data sets is computationally e...
The Metropolis-adjusted Langevin (MALA) algorithm is a sampling algorithm which makes local moves by...
This thesis focuses on adaptive Stochastic Gradient Langevin Dynamics (SGLD) algorithms to solve opt...
Stochastic gradient MCMC methods, such as stochastic gradient Langevin dynamics (SGLD), employ fast ...
In this paper, we study the computational complexity of sampling from a Bayesian posterior (or pseud...