Funder: Gates Cambridge Trust (GB)AbstractVariance reduction is a crucial tool for improving the slow convergence of stochastic gradient descent. Only a few variance-reduced methods, however, have yet been shown to directly benefit from Nesterov’s acceleration techniques to match the convergence rates of accelerated gradient methods. Such approaches rely on “negative momentum”, a technique for further variance reduction that is generally specific to the SVRG gradient estimator. In this work, we show for the first time that negative momentum is unnecessary for acceleration and develop a universal acceleration framework that allows all popular variance-reduced methods to achieve accelerated convergence rates. The constants appearing in these ...
The article examines in some detail the convergence rate and mean-square-error performance of moment...
In this paper, we study a stochastic strongly convex optimization problem and propose three classes ...
Stochastic Gradient Descent (SGD) has played a crucial role in the success of modern machine learnin...
Variance reduction is a crucial tool for improving the slow convergence of stochastic gradient desce...
© 1989-2012 IEEE. In this paper, we propose a simple variant of the original SVRG, called variance r...
Stochastic gradient descent is popular for large scale optimization but has slow convergence asympto...
We study the convergence of accelerated stochastic gradient descent for strongly convex objectives u...
Stochastic gradient descent is popular for large scale optimization but has slow convergence asympto...
With the purpose of examining biased updates in variance-reduced stochastic gradient methods, we int...
Recently, many variance reduced stochastic alternating direction method of multipliers (ADMM) method...
Stochastic variance reduced gradient (SVRG) is a popular variance reduction technique for accelerati...
International audienceIn this paper, we propose a novel reinforcement-learning algorithm consisting ...
International audienceIn this paper, we propose a unified view of gradient-based algorithms for stoc...
The stochastic gradient Langevin Dynamics is one of the most fundamental algorithms to solve samplin...
With the purpose of examining biased updates in variance-reduced stochastic gradient methods, we int...
The article examines in some detail the convergence rate and mean-square-error performance of moment...
In this paper, we study a stochastic strongly convex optimization problem and propose three classes ...
Stochastic Gradient Descent (SGD) has played a crucial role in the success of modern machine learnin...
Variance reduction is a crucial tool for improving the slow convergence of stochastic gradient desce...
© 1989-2012 IEEE. In this paper, we propose a simple variant of the original SVRG, called variance r...
Stochastic gradient descent is popular for large scale optimization but has slow convergence asympto...
We study the convergence of accelerated stochastic gradient descent for strongly convex objectives u...
Stochastic gradient descent is popular for large scale optimization but has slow convergence asympto...
With the purpose of examining biased updates in variance-reduced stochastic gradient methods, we int...
Recently, many variance reduced stochastic alternating direction method of multipliers (ADMM) method...
Stochastic variance reduced gradient (SVRG) is a popular variance reduction technique for accelerati...
International audienceIn this paper, we propose a novel reinforcement-learning algorithm consisting ...
International audienceIn this paper, we propose a unified view of gradient-based algorithms for stoc...
The stochastic gradient Langevin Dynamics is one of the most fundamental algorithms to solve samplin...
With the purpose of examining biased updates in variance-reduced stochastic gradient methods, we int...
The article examines in some detail the convergence rate and mean-square-error performance of moment...
In this paper, we study a stochastic strongly convex optimization problem and propose three classes ...
Stochastic Gradient Descent (SGD) has played a crucial role in the success of modern machine learnin...