The analysis of gradient descent-type methods typically relies on the Lipschitz continuity of the objective gradient. This generally requires an expensive hyperparameter tuning process to appropriately calibrate a stepsize for a given problem. In this work we introduce a local first-order smoothness oracle (LFSO) which generalizes the Lipschitz continuous gradients smoothness condition and is applicable to any twice-differentiable function. We show that this oracle can encode all relevant problem information for tuning stepsizes for a suitably modified gradient descent method and give global and local convergence results. We also show that LFSOs in this modified first-order method can yield global linear convergence rates for non-strongly c...
We present and computationally evaluate a variant of the fast gradient method by Nesterov that is ca...
We present and computationally evaluate a variant of the fast gradient method by Nesterov that is ca...
International audienceWe propose a new family of adaptive first-order methods for a class of convex ...
Classical global convergence results for first-order methods rely on uniform smoothness and the \L{}...
Several important problems in learning theory and data science involve high-dimensional optimization...
Several important problems in learning theory and data science involve high-dimensional optimization...
Several important problems in learning theory and data science involve high-dimensional optimization...
Several important problems in learning theory and data science involve high-dimensional optimization...
Several important problems in learning theory and data science involve high-dimensional optimization...
2013-2014 > Academic research: refereed > Publication in refereed journalVersion of RecordPublishe
We present a strikingly simple proof that two rules are sufficient to automate gradient descent: 1) ...
Polyak-{\L}ojasiewicz (PL) [Polyak, 1963] condition is a weaker condition than the strong convexity ...
The usual approach to developing and analyzing first-order methods for smooth convex optimization as...
We present and computationally evaluate a variant of the fast gradient method by Nesterov that is ca...
We propose a new first-order method for minimizing nonconvex functions with a Lipschitz continuous g...
We present and computationally evaluate a variant of the fast gradient method by Nesterov that is ca...
We present and computationally evaluate a variant of the fast gradient method by Nesterov that is ca...
International audienceWe propose a new family of adaptive first-order methods for a class of convex ...
Classical global convergence results for first-order methods rely on uniform smoothness and the \L{}...
Several important problems in learning theory and data science involve high-dimensional optimization...
Several important problems in learning theory and data science involve high-dimensional optimization...
Several important problems in learning theory and data science involve high-dimensional optimization...
Several important problems in learning theory and data science involve high-dimensional optimization...
Several important problems in learning theory and data science involve high-dimensional optimization...
2013-2014 > Academic research: refereed > Publication in refereed journalVersion of RecordPublishe
We present a strikingly simple proof that two rules are sufficient to automate gradient descent: 1) ...
Polyak-{\L}ojasiewicz (PL) [Polyak, 1963] condition is a weaker condition than the strong convexity ...
The usual approach to developing and analyzing first-order methods for smooth convex optimization as...
We present and computationally evaluate a variant of the fast gradient method by Nesterov that is ca...
We propose a new first-order method for minimizing nonconvex functions with a Lipschitz continuous g...
We present and computationally evaluate a variant of the fast gradient method by Nesterov that is ca...
We present and computationally evaluate a variant of the fast gradient method by Nesterov that is ca...
International audienceWe propose a new family of adaptive first-order methods for a class of convex ...