International audienceStochastic optimization algorithms with variance reduction have proven successful for minimizing large finite sums of functions. Unfortunately, these techniques are unable to deal with stochastic perturbations of input data, induced for example by data augmentation. In such cases, the objective is no longer a finite sum, and the main candidate for optimization is the stochastic gradient descent method (SGD). In this paper, we introduce a variance reduction approach for these settings when the objective is composite and strongly convex. The convergence rate outperforms SGD with a typically much smaller constant factor, which depends on the variance of gradient estimates only due to perturbations on a single example
Stochastic gradient descent (SGD) is a simple and popular method to solve stochastic optimization pr...
International audienceWe consider the problem of training machine learning models on distributed dat...
In this thesis, we discuss and develop randomized algorithms for big data problems. In particular, w...
International audienceIn this paper, we propose a unified view of gradient-based algorithms for stoc...
The notable changes over the current version: - worked example of convergence rates showing SAG can ...
Revision from January 2015 submission. Major changes: updated literature follow and discussion of su...
© 1989-2012 IEEE. In this paper, we propose a simple variant of the original SVRG, called variance r...
In this paper, we develop stochastic variance reduced algorithms for solving a class of finite-sum m...
International audienceWe consider convex-concave saddle-point problems where the objective functions...
International audienceIn this paper, we introduce various mechanisms to obtain accelerated first-ord...
Stochastic optimization has received extensive attention in recent years due to their extremely pote...
Variational inequalities are a broad formalism that encompasses a vast number of applications. Motiv...
<p>Stochastic gradient optimization is a class of widely used algorithms for training machine learni...
Abstract We consider the problem of minimizing the sum of two convex functions: one is the average o...
We study optimization algorithms for the finite sum problems frequently arising in machine learning...
Stochastic gradient descent (SGD) is a simple and popular method to solve stochastic optimization pr...
International audienceWe consider the problem of training machine learning models on distributed dat...
In this thesis, we discuss and develop randomized algorithms for big data problems. In particular, w...
International audienceIn this paper, we propose a unified view of gradient-based algorithms for stoc...
The notable changes over the current version: - worked example of convergence rates showing SAG can ...
Revision from January 2015 submission. Major changes: updated literature follow and discussion of su...
© 1989-2012 IEEE. In this paper, we propose a simple variant of the original SVRG, called variance r...
In this paper, we develop stochastic variance reduced algorithms for solving a class of finite-sum m...
International audienceWe consider convex-concave saddle-point problems where the objective functions...
International audienceIn this paper, we introduce various mechanisms to obtain accelerated first-ord...
Stochastic optimization has received extensive attention in recent years due to their extremely pote...
Variational inequalities are a broad formalism that encompasses a vast number of applications. Motiv...
<p>Stochastic gradient optimization is a class of widely used algorithms for training machine learni...
Abstract We consider the problem of minimizing the sum of two convex functions: one is the average o...
We study optimization algorithms for the finite sum problems frequently arising in machine learning...
Stochastic gradient descent (SGD) is a simple and popular method to solve stochastic optimization pr...
International audienceWe consider the problem of training machine learning models on distributed dat...
In this thesis, we discuss and develop randomized algorithms for big data problems. In particular, w...