We propose a new first-order method for minimizing nonconvex functions with a Lipschitz continuous gradient and Hessian. The proposed method is an accelerated gradient descent with two restart mechanisms and finds a solution where the gradient norm is less than $\epsilon$ in $O(\epsilon^{-7/4})$ function and gradient evaluations. Unlike existing algorithms with similar complexity bounds, our method is parameter-free because it requires no prior knowledge of problem-dependent parameters, e.g., the Lipschitz constants and the target accuracy $\epsilon$. The main challenge in achieving this advantage is estimating the Lipschitz constant of the Hessian using only first-order information. To this end, we develop a new Hessian-free analysis based...
In this paper, we propose an interior-point method for linearly constrained optimization problems (p...
Several important problems in learning theory and data science involve high-dimensional optimization...
AbstractIn this paper the development, convergence theory and numerical testing of a class of gradie...
Nonconvex optimization with great demand of fast solvers is ubiquitous in modern machine learning. T...
International audienceWe propose a new family of adaptive first-order methods for a class of convex ...
International audienceIn this paper, we propose an interior-point method for linearly constrained-an...
International audienceIn this paper, we propose an interior-point method for linearly constrained-an...
The analysis of gradient descent-type methods typically relies on the Lipschitz continuity of the ob...
International audienceWe introduce a generic scheme to solve non-convex optimization problems using ...
Several important problems in learning theory and data science involve high-dimensional optimization...
International audienceWe introduce a generic scheme to solve non-convex optimization problems using ...
Several important problems in learning theory and data science involve high-dimensional optimization...
In this work, we develop first-order (Hessian-free) and zero-order (derivative-free) implementations...
Several important problems in learning theory and data science involve high-dimensional optimization...
Several important problems in learning theory and data science involve high-dimensional optimization...
In this paper, we propose an interior-point method for linearly constrained optimization problems (p...
Several important problems in learning theory and data science involve high-dimensional optimization...
AbstractIn this paper the development, convergence theory and numerical testing of a class of gradie...
Nonconvex optimization with great demand of fast solvers is ubiquitous in modern machine learning. T...
International audienceWe propose a new family of adaptive first-order methods for a class of convex ...
International audienceIn this paper, we propose an interior-point method for linearly constrained-an...
International audienceIn this paper, we propose an interior-point method for linearly constrained-an...
The analysis of gradient descent-type methods typically relies on the Lipschitz continuity of the ob...
International audienceWe introduce a generic scheme to solve non-convex optimization problems using ...
Several important problems in learning theory and data science involve high-dimensional optimization...
International audienceWe introduce a generic scheme to solve non-convex optimization problems using ...
Several important problems in learning theory and data science involve high-dimensional optimization...
In this work, we develop first-order (Hessian-free) and zero-order (derivative-free) implementations...
Several important problems in learning theory and data science involve high-dimensional optimization...
Several important problems in learning theory and data science involve high-dimensional optimization...
In this paper, we propose an interior-point method for linearly constrained optimization problems (p...
Several important problems in learning theory and data science involve high-dimensional optimization...
AbstractIn this paper the development, convergence theory and numerical testing of a class of gradie...