In this work, we show that the heavy-ball ($\HB$) method provably does not reach an accelerated convergence rate on smooth strongly convex problems. More specifically, we show that for any condition number and any choice of algorithmic parameters, either the worst-case convergence rate of $\HB$ on the class of $L$-smooth and $μ$-strongly convex \textit{quadratic} functions is not accelerated (that is, slower than $1 - \mathcal{O}(κ)$), or there exists an $L$-smooth $μ$-strongly convex function and an initialization such that the method does not converge. To the best of our knowledge, this result closes a simple yet open question on one of the most used and iconic first-order optimization technique. Our approach builds on finding functions f...
Nonconvex optimization with great demand of fast solvers is ubiquitous in modern machine learning. T...
The goal of this paper is to study the effect of inexact first-order information on the first-order ...
We study stochastic gradient descent (SGD) and the stochastic heavy ball method (SHB, otherwise know...
In this paper, we study the behavior of solutions of the ODE associated to the Heavy Ball method. Si...
In this paper, a joint study of the behavior of solutions of the Heavy Ball ODE and Heavy Ball type ...
In this paper, we revisit the convergence of the Heavy-ball method, and present improved convergence...
This paper deals with a natural stochastic optimization procedure derived from the so-called Heavy-b...
International audienceThis paper deals with a natural stochastic optimization procedure derived from...
We analyze worst-case convergence guarantees of first-order optimization methods over a function cla...
Heavy Ball (HB) nowadays is one of the most popular momentum methods in non-convex optimization. It ...
In this paper, we analyze the convergence rate of the Heavy-ball algorithm applied to optimize a cla...
In a Hilbertian framework, for the minimization of a general convex differentiable function f , we i...
We show that accelerated gradient descent, averaged gradient descent and the heavy-ball method for q...
The standard assumption for proving linear convergence of first order methods for smooth convex opti...
We show that the exact worst-case performance of fixed-step first-order methods for smooth (possibly...
Nonconvex optimization with great demand of fast solvers is ubiquitous in modern machine learning. T...
The goal of this paper is to study the effect of inexact first-order information on the first-order ...
We study stochastic gradient descent (SGD) and the stochastic heavy ball method (SHB, otherwise know...
In this paper, we study the behavior of solutions of the ODE associated to the Heavy Ball method. Si...
In this paper, a joint study of the behavior of solutions of the Heavy Ball ODE and Heavy Ball type ...
In this paper, we revisit the convergence of the Heavy-ball method, and present improved convergence...
This paper deals with a natural stochastic optimization procedure derived from the so-called Heavy-b...
International audienceThis paper deals with a natural stochastic optimization procedure derived from...
We analyze worst-case convergence guarantees of first-order optimization methods over a function cla...
Heavy Ball (HB) nowadays is one of the most popular momentum methods in non-convex optimization. It ...
In this paper, we analyze the convergence rate of the Heavy-ball algorithm applied to optimize a cla...
In a Hilbertian framework, for the minimization of a general convex differentiable function f , we i...
We show that accelerated gradient descent, averaged gradient descent and the heavy-ball method for q...
The standard assumption for proving linear convergence of first order methods for smooth convex opti...
We show that the exact worst-case performance of fixed-step first-order methods for smooth (possibly...
Nonconvex optimization with great demand of fast solvers is ubiquitous in modern machine learning. T...
The goal of this paper is to study the effect of inexact first-order information on the first-order ...
We study stochastic gradient descent (SGD) and the stochastic heavy ball method (SHB, otherwise know...