The notable changes over the current version: - worked example of convergence rates showing SAG can be faster than first-order methods - pointing out that the storage cost is O(n) for linear models - the more-stable line-search - comparison to additional optimal SG methods - comparison to rates of coordinate descent methods in quadratic case.International audienceWe propose a new stochastic gradient method for optimizing the sum of a finite set of smooth functions, where the sum is strongly convex. While standard stochastic gradient methods converge at sublinear rates for this problem, the proposed method incorporates a memory of previous gradient values in order to achieve a linear convergence rate. In a machine learning context, numerical...
Stochastic gradient descent is the method of choice for solving large-scale optimization problems in...
International audienceWe consider the stochastic approximation problem where a convex function has t...
AbstractIn this paper, a stochastic gradient descent algorithm is proposed for the binary classifica...
Revision from January 2015 submission. Major changes: updated literature follow and discussion of su...
Despite the recent growth of theoretical studies and empirical successes of neural networks, gradien...
Stochastic gradient descent (SGD) is a simple and popular method to solve stochastic optimization pr...
We propose the stochastic average gradient (SAG) method for optimizing the sum of a finite number of...
Stochastic gradient descent (SGD) is a simple and popular method to solve stochastic optimization pr...
We propose the stochastic average gradient (SAG) method for optimizing the sum of a finite number of...
We aim to make stochastic gradient descent (SGD) adaptive to (i) the noise $\sigma^2$ in the stochas...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
Stochastic gradient descent (SGD) is a sim-ple and popular method to solve stochas-tic optimization ...
International audienceWe consider binary classification problems with positive definite kernels and ...
We design step-size schemes that make stochastic gradient descent (SGD) adaptive to (i) the noise σ ...
Stochastic gradient descent is the method of choice for solving large-scale optimization problems in...
International audienceWe consider the stochastic approximation problem where a convex function has t...
AbstractIn this paper, a stochastic gradient descent algorithm is proposed for the binary classifica...
Revision from January 2015 submission. Major changes: updated literature follow and discussion of su...
Despite the recent growth of theoretical studies and empirical successes of neural networks, gradien...
Stochastic gradient descent (SGD) is a simple and popular method to solve stochastic optimization pr...
We propose the stochastic average gradient (SAG) method for optimizing the sum of a finite number of...
Stochastic gradient descent (SGD) is a simple and popular method to solve stochastic optimization pr...
We propose the stochastic average gradient (SAG) method for optimizing the sum of a finite number of...
We aim to make stochastic gradient descent (SGD) adaptive to (i) the noise $\sigma^2$ in the stochas...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
Stochastic gradient descent (SGD) is a sim-ple and popular method to solve stochas-tic optimization ...
International audienceWe consider binary classification problems with positive definite kernels and ...
We design step-size schemes that make stochastic gradient descent (SGD) adaptive to (i) the noise σ ...
Stochastic gradient descent is the method of choice for solving large-scale optimization problems in...
International audienceWe consider the stochastic approximation problem where a convex function has t...
AbstractIn this paper, a stochastic gradient descent algorithm is proposed for the binary classifica...