Nesterov's Accelerated Gradient (NAG) for optimization has better performance than its continuous time limit (noiseless kinetic Langevin) when a finite step-size is employed \citep{shi2021understanding}. This work explores the sampling counterpart of this phenonemon and proposes a diffusion process, whose discretizations can yield accelerated gradient-based MCMC methods. More precisely, we reformulate the optimizer of NAG for strongly convex functions (NAG-SC) as a Hessian-Free High-Resolution ODE, change its high-resolution coefficient to a hyperparameter, inject appropriate noise, and discretize the resulting diffusion process. The acceleration effect of the new hyperparameter is quantified and it is not an artificial one created by time-...
Sampling from probability distributions is a problem of significant importance in Statistics and Mac...
We aim to make stochastic gradient descent (SGD) adaptive to (i) the noise $\sigma^2$ in the stochas...
We consider the problem of sampling from a distribution governed by a potential function. This work ...
We formulate gradient-based Markov chain Monte Carlo (MCMC) sampling as optimization on the space of...
Langevin dynamics-based sampling algorithms are arguably among the most widelyused Markov Chain Mont...
We study the convergence of accelerated stochastic gradient descent for strongly convex objectives u...
In the history of first-order algorithms, Nesterov's accelerated gradient descent (NAG) is one of th...
We provide a novel accelerated first-order method that achieves the asymptotically optimal convergen...
We provide a novel accelerated first-order method that achieves the asymptotically optimal convergen...
International audienceWe show that accelerated gradient descent, averaged gradient descent and the h...
International audienceWe consider the stochastic optimization problem where a convex function is min...
The Metropolis-adjusted Langevin (MALA) algorithm is a sampling algorithm which makes local moves by...
Acceleration in optimization is a term that is generally applied to optimization algorithms presenti...
The dissertation addresses the research topics of machine learning outlined below. We developed the ...
We investigate convex differentiable optimization and explore the temporal discretization of damped ...
Sampling from probability distributions is a problem of significant importance in Statistics and Mac...
We aim to make stochastic gradient descent (SGD) adaptive to (i) the noise $\sigma^2$ in the stochas...
We consider the problem of sampling from a distribution governed by a potential function. This work ...
We formulate gradient-based Markov chain Monte Carlo (MCMC) sampling as optimization on the space of...
Langevin dynamics-based sampling algorithms are arguably among the most widelyused Markov Chain Mont...
We study the convergence of accelerated stochastic gradient descent for strongly convex objectives u...
In the history of first-order algorithms, Nesterov's accelerated gradient descent (NAG) is one of th...
We provide a novel accelerated first-order method that achieves the asymptotically optimal convergen...
We provide a novel accelerated first-order method that achieves the asymptotically optimal convergen...
International audienceWe show that accelerated gradient descent, averaged gradient descent and the h...
International audienceWe consider the stochastic optimization problem where a convex function is min...
The Metropolis-adjusted Langevin (MALA) algorithm is a sampling algorithm which makes local moves by...
Acceleration in optimization is a term that is generally applied to optimization algorithms presenti...
The dissertation addresses the research topics of machine learning outlined below. We developed the ...
We investigate convex differentiable optimization and explore the temporal discretization of damped ...
Sampling from probability distributions is a problem of significant importance in Statistics and Mac...
We aim to make stochastic gradient descent (SGD) adaptive to (i) the noise $\sigma^2$ in the stochas...
We consider the problem of sampling from a distribution governed by a potential function. This work ...