We consider the optimization of a quadratic objective function whose gradients are only accessible through a stochastic oracle that returns the gradient at any given point plus a zero-mean finite variance random error. We present the first algorithm that achieves jointly the optimal prediction error rates for least-squares regression, both in terms of forgetting the initial conditions in O (1 / n(2)), and in terms of dependence on the noise and dimension d of the problem, as O (d / n). Our new algorithm is based on averaged accelerated regularized gradient descent, and may also be analyzed through finer assumptions on initial conditions and the Hessian matrix, leading to dimension-free quantities that may still be small in some distances wh...
Many problems in machine learning are naturally cast as the minimization of a smooth function define...
We consider learning methods based on the regularization of a convex empirical risk by a squared Hil...
We study nonlinear least-squares problem that can be transformed to linear problem by change of vari...
International audienceWe consider the optimization of a quadratic objective function whose gradients...
International audienceWe consider the optimization of a quadratic objective function whose gradients...
International audienceWe consider the optimization of a quadratic objective function whose gradients...
We investigate implicit regularization schemes for gradient descent methods applied to unpenalized l...
In this paper, we consider an online least square regression problem where the objective function is...
We investigate implicit regularization schemes for gradient descent methods applied to unpenalized l...
Online learning algorithms require to often recompute least squares regression estimates of paramete...
Online learning algorithms require to often recompute least squares regression estimates of paramete...
International audienceIn this paper, we investigate the impact of compression on stochastic gradient...
International audienceWe consider the random-design least-squares regression problem within the repr...
Many problems in machine learning are naturally cast as the minimization of a smooth function define...
Many problems in machine learning are naturally cast as the minimization of a smooth function define...
Many problems in machine learning are naturally cast as the minimization of a smooth function define...
We consider learning methods based on the regularization of a convex empirical risk by a squared Hil...
We study nonlinear least-squares problem that can be transformed to linear problem by change of vari...
International audienceWe consider the optimization of a quadratic objective function whose gradients...
International audienceWe consider the optimization of a quadratic objective function whose gradients...
International audienceWe consider the optimization of a quadratic objective function whose gradients...
We investigate implicit regularization schemes for gradient descent methods applied to unpenalized l...
In this paper, we consider an online least square regression problem where the objective function is...
We investigate implicit regularization schemes for gradient descent methods applied to unpenalized l...
Online learning algorithms require to often recompute least squares regression estimates of paramete...
Online learning algorithms require to often recompute least squares regression estimates of paramete...
International audienceIn this paper, we investigate the impact of compression on stochastic gradient...
International audienceWe consider the random-design least-squares regression problem within the repr...
Many problems in machine learning are naturally cast as the minimization of a smooth function define...
Many problems in machine learning are naturally cast as the minimization of a smooth function define...
Many problems in machine learning are naturally cast as the minimization of a smooth function define...
We consider learning methods based on the regularization of a convex empirical risk by a squared Hil...
We study nonlinear least-squares problem that can be transformed to linear problem by change of vari...