We perform the first tight convergence analysis of the gradient method with varying step sizes when applied to smooth hypoconvex (weakly convex) functions. Hypoconvex functions are smooth nonconvex functions whose curvature is bounded and assumed to belong to the interval $[\mu, L]$, with $\mu<0$. Our convergence rates improve and extend the existing analysis for smooth nonconvex functions with $L$-Lipschitz gradient (which corresponds to the case $\mu=-L$), and smoothly interpolates between that class and the class of smooth convex functions. We obtain our results using the performance estimation framework adapted to hypoconvex functions, for which new interpolation conditions are derived. We derive explicit upper bounds on the minimum gra...
International audienceIn this paper we provide a theoretical and numerical comparison of convergence...
The analysis of gradient descent-type methods typically relies on the Lipschitz continuity of the ob...
6 pagesWe give a simple proof that the Frank-Wolfe algorithm obtains a stationary point at a rate of...
The convergence behavior of gradient methods for minimizing convex differentiable functions is one o...
The standard assumption for proving linear convergence of first order methods for smooth convex opti...
We show that the basic stochastic gradient method applied to a strongly-convex differentiable functi...
We extend the previous analysis of Schmidt et al. [2011] to derive the linear convergence rate obtai...
We consider the gradient (or steepest) descent method with exact line search applied to a strongly c...
The study of first-order optimization is sensitive to the assumptions made on the objective function...
We consider the gradient (or steepest) descent method with exact line search applied to a strongly c...
International audienceThis paper studies the asymptotic behavior of the constant step Stochastic Gra...
We consider the gradient (or steepest) descent method with exact line search applied to a strongly c...
Based on a result by Taylor et al. (J Optim Theory Appl 178(2):455–476, 2018) on the attainable conv...
Analysis of the convergence rates of modern convex optimization algorithms can be achived through bi...
We consider the gradient (or steepest) descent method with exact line search applied to a strongly c...
International audienceIn this paper we provide a theoretical and numerical comparison of convergence...
The analysis of gradient descent-type methods typically relies on the Lipschitz continuity of the ob...
6 pagesWe give a simple proof that the Frank-Wolfe algorithm obtains a stationary point at a rate of...
The convergence behavior of gradient methods for minimizing convex differentiable functions is one o...
The standard assumption for proving linear convergence of first order methods for smooth convex opti...
We show that the basic stochastic gradient method applied to a strongly-convex differentiable functi...
We extend the previous analysis of Schmidt et al. [2011] to derive the linear convergence rate obtai...
We consider the gradient (or steepest) descent method with exact line search applied to a strongly c...
The study of first-order optimization is sensitive to the assumptions made on the objective function...
We consider the gradient (or steepest) descent method with exact line search applied to a strongly c...
International audienceThis paper studies the asymptotic behavior of the constant step Stochastic Gra...
We consider the gradient (or steepest) descent method with exact line search applied to a strongly c...
Based on a result by Taylor et al. (J Optim Theory Appl 178(2):455–476, 2018) on the attainable conv...
Analysis of the convergence rates of modern convex optimization algorithms can be achived through bi...
We consider the gradient (or steepest) descent method with exact line search applied to a strongly c...
International audienceIn this paper we provide a theoretical and numerical comparison of convergence...
The analysis of gradient descent-type methods typically relies on the Lipschitz continuity of the ob...
6 pagesWe give a simple proof that the Frank-Wolfe algorithm obtains a stationary point at a rate of...