We consider the gradient (or steepest) descent method with exact line search applied to a strongly convex function with Lipschitz continuous gradient. We establish the exact worst-case rate of convergence of this scheme, and show that this worst-case behavior is exhibited by a certain convex quadratic function. We also extend the result to a noisy variant of gradient descent method, where exact line-search is performed in a search direction that differs from negative gradient by at most a prescribed relative tolerance. The proof is computer-assisted, and relies on the resolution of semidefinite programming performance estimation problems as introduced in the paper [Y. Drori and M. Teboulle. Performance of first-order methods for smooth conv...
We provide new tools for worst-case performance analysis of the gradient (or steepest descent) metho...
We analyze worst-case convergence guarantees of first-order optimization methods over a function cla...
We analyze worst-case convergence guarantees of first-order optimization methods over a function cla...
We consider the gradient (or steepest) descent method with exact line search applied to a strongly c...
We consider the gradient (or steepest) descent method with exact line search applied to a strongly c...
We consider the gradient (or steepest) descent method with exact line search applied to a strongly c...
We consider the gradient (or steepest) descent method with exact line search applied to a strongly c...
In this paper we prove that the broad class of direct-search methods of directional type, based on i...
In this paper we prove that the broad class of direct-search methods of directional type, based on i...
The worst-case complexity of the steepest-descent algorithm with exact line-searches for unconstrain...
The convergence behavior of gradient methods for minimizing convex differentiable functions is one o...
We show that the exact worst-case performance of fixed-step first-order methods for smooth (possibly...
We study the worst-case convergence rates of the proximal gradient method for minimizing the sum of ...
We provide new tools for worst-case performance analysis of the gradient (or steepest descent) metho...
We provide new tools for worst-case performance analysis of the gradient (or steepest descent) metho...
We provide new tools for worst-case performance analysis of the gradient (or steepest descent) metho...
We analyze worst-case convergence guarantees of first-order optimization methods over a function cla...
We analyze worst-case convergence guarantees of first-order optimization methods over a function cla...
We consider the gradient (or steepest) descent method with exact line search applied to a strongly c...
We consider the gradient (or steepest) descent method with exact line search applied to a strongly c...
We consider the gradient (or steepest) descent method with exact line search applied to a strongly c...
We consider the gradient (or steepest) descent method with exact line search applied to a strongly c...
In this paper we prove that the broad class of direct-search methods of directional type, based on i...
In this paper we prove that the broad class of direct-search methods of directional type, based on i...
The worst-case complexity of the steepest-descent algorithm with exact line-searches for unconstrain...
The convergence behavior of gradient methods for minimizing convex differentiable functions is one o...
We show that the exact worst-case performance of fixed-step first-order methods for smooth (possibly...
We study the worst-case convergence rates of the proximal gradient method for minimizing the sum of ...
We provide new tools for worst-case performance analysis of the gradient (or steepest descent) metho...
We provide new tools for worst-case performance analysis of the gradient (or steepest descent) metho...
We provide new tools for worst-case performance analysis of the gradient (or steepest descent) metho...
We analyze worst-case convergence guarantees of first-order optimization methods over a function cla...
We analyze worst-case convergence guarantees of first-order optimization methods over a function cla...