The vast majority of convergence rates analysis for stochastic gradient methods in the literature focus on convergence in expectation, whereas trajectory-wise almost sure convergence is clearly important to ensure that any instantiation of the stochastic algorithms would converge with probability one. Here we provide a unified almost sure convergence rates analysis for stochastic gradient descent (SGD), stochastic heavy-ball (SHB), and stochastic Nesterov's accelerated gradient (SNAG) methods. We show, for the first time, that the almost sure convergence rates obtained for these stochastic gradient methods on strongly convex functions, are arbitrarily close to their optimal convergence rates possible. For non-convex objective functions, we ...
We show that the basic stochastic gradient method applied to a strongly-convex differentiable functi...
Stochastic Gradient Descent (SGD) is the workhorse beneath the deep learning revolution. However, SG...
Thesis (Ph.D.)--University of Washington, 2019Tremendous advances in large scale machine learning an...
We study stochastic gradient descent (SGD) and the stochastic heavy ball method (SHB, otherwise know...
An usual problem in statistics consists in estimating the minimizer of a convex function. When we ha...
Stochastic gradient descent (SGD) is a simple and popular method to solve stochastic optimization pr...
An usual problem in statistics consists in estimating the minimizer of a convex function. When we ha...
With a weighting scheme proportional to t, a traditional stochastic gradient descent (SGD) algorithm...
Consider the problem of minimizing functions that are Lipschitz and strongly convex, but not necessa...
Stochastic Gradient Descent (SGD) is one of the simplest and most popular stochastic optimization me...
Stochastic gradient descent (SGD) is a sim-ple and popular method to solve stochas-tic optimization ...
International audienceRecent studies have provided both empirical and theoretical evidence illustrat...
Abstract Stochastic gradient descent algorithm is a classical and useful method for stochastic optim...
We prove the convergence to minima and estimates on the rate of convergence for the stochastic gradi...
Despite the recent growth of theoretical studies and empirical successes of neural networks, gradien...
We show that the basic stochastic gradient method applied to a strongly-convex differentiable functi...
Stochastic Gradient Descent (SGD) is the workhorse beneath the deep learning revolution. However, SG...
Thesis (Ph.D.)--University of Washington, 2019Tremendous advances in large scale machine learning an...
We study stochastic gradient descent (SGD) and the stochastic heavy ball method (SHB, otherwise know...
An usual problem in statistics consists in estimating the minimizer of a convex function. When we ha...
Stochastic gradient descent (SGD) is a simple and popular method to solve stochastic optimization pr...
An usual problem in statistics consists in estimating the minimizer of a convex function. When we ha...
With a weighting scheme proportional to t, a traditional stochastic gradient descent (SGD) algorithm...
Consider the problem of minimizing functions that are Lipschitz and strongly convex, but not necessa...
Stochastic Gradient Descent (SGD) is one of the simplest and most popular stochastic optimization me...
Stochastic gradient descent (SGD) is a sim-ple and popular method to solve stochas-tic optimization ...
International audienceRecent studies have provided both empirical and theoretical evidence illustrat...
Abstract Stochastic gradient descent algorithm is a classical and useful method for stochastic optim...
We prove the convergence to minima and estimates on the rate of convergence for the stochastic gradi...
Despite the recent growth of theoretical studies and empirical successes of neural networks, gradien...
We show that the basic stochastic gradient method applied to a strongly-convex differentiable functi...
Stochastic Gradient Descent (SGD) is the workhorse beneath the deep learning revolution. However, SG...
Thesis (Ph.D.)--University of Washington, 2019Tremendous advances in large scale machine learning an...