We consider the optimization of a smooth and strongly convex objective using constant step-size stochastic gradient descent (SGD) and study its properties through the prism of Markov chains. We show that, for unbiased gradient estimates with mildly controlled variance, the iteration converges to an invariant distribution in total variation distance. We also establish this convergence in Wasserstein-2 distance in a more general setting compared to previous work. Thanks to the invariance property of the limit distribution, our analysis shows that the latter inherits sub-Gaussian or sub-exponential concentration properties when these hold true for the gradient. This allows the derivation of high-confidence bounds for the final estimate. Finall...
Abstract: Stochastic gradient descent is an optimisation method that combines classical gradient des...
We consider optimizing a function smooth convex function $f$ that is the average of a set of differe...
International audienceWe consider the least-squares regression problem and provide a detailed asympt...
We consider the minimization of an objective function given access to unbiased estimates of its grad...
We establish a sharp uniform-in-time error estimate for the Stochastic Gradient Langevin Dynamics (S...
International audienceRecent studies have provided both empirical and theoretical evidence illustrat...
Recently, Loizou et al. (2021), proposed and analyzed stochastic gradient descent (SGD) with stochas...
Recently there is a large amount of work devoted to the study of Markov chain stochastic gradient me...
The strong growth condition (SGC) is known to be a sufficient condition for linear convergence of th...
94 pages, 4 figuresThis paper proposes a thorough theoretical analysis of Stochastic Gradient Descen...
We design step-size schemes that make stochastic gradient descent (SGD) adaptive to (i) the noise σ ...
In this thesis, Stochastic Gradient Descent (SGD), an optimization method originally popular due to ...
We aim to make stochastic gradient descent (SGD) adaptive to (i) the noise $\sigma^2$ in the stochas...
We consider Linear Stochastic Approximation (LSA) with a constant stepsize and Markovian data. Viewi...
In this paper, we propose a new covering technique localized for the trajectories of SGD. This local...
Abstract: Stochastic gradient descent is an optimisation method that combines classical gradient des...
We consider optimizing a function smooth convex function $f$ that is the average of a set of differe...
International audienceWe consider the least-squares regression problem and provide a detailed asympt...
We consider the minimization of an objective function given access to unbiased estimates of its grad...
We establish a sharp uniform-in-time error estimate for the Stochastic Gradient Langevin Dynamics (S...
International audienceRecent studies have provided both empirical and theoretical evidence illustrat...
Recently, Loizou et al. (2021), proposed and analyzed stochastic gradient descent (SGD) with stochas...
Recently there is a large amount of work devoted to the study of Markov chain stochastic gradient me...
The strong growth condition (SGC) is known to be a sufficient condition for linear convergence of th...
94 pages, 4 figuresThis paper proposes a thorough theoretical analysis of Stochastic Gradient Descen...
We design step-size schemes that make stochastic gradient descent (SGD) adaptive to (i) the noise σ ...
In this thesis, Stochastic Gradient Descent (SGD), an optimization method originally popular due to ...
We aim to make stochastic gradient descent (SGD) adaptive to (i) the noise $\sigma^2$ in the stochas...
We consider Linear Stochastic Approximation (LSA) with a constant stepsize and Markovian data. Viewi...
In this paper, we propose a new covering technique localized for the trajectories of SGD. This local...
Abstract: Stochastic gradient descent is an optimisation method that combines classical gradient des...
We consider optimizing a function smooth convex function $f$ that is the average of a set of differe...
International audienceWe consider the least-squares regression problem and provide a detailed asympt...