The move from hand-designed to learned optimizers in machine learning has been quite successful for gradient-based and -free optimizers. When facing a constrained problem, however, maintaining feasibility typically requires a projection step, which might be computationally expensive and not differentiable. We show how the design of projection-free convex optimization algorithms can be cast as a learning problem based on Frank-Wolfe Networks: recurrent networks implementing the Frank-Wolfe algorithm aka. conditional gradients. This allows them to learn to exploit structure when, e.g., optimizing over rank-1 matrices. Our LSTM-learned optimizers outperform hand-designed as well learned but unconstrained ones. We demonstrate this for training ...
International audienceIn a series of recent theoretical works, it was shown that strongly over-param...
Aiming at convex optimization under structural constraints, this work introduces and analyzes a vari...
Large-scale machine learning problems can be reduced to non-convex optimization problems if state-of...
We revisit the Frank-Wolfe algorithm for constrained convex optimization and show that it can be imp...
Learning a deep neural network requires solving a challenging optimization problem: it is a high-dim...
The Frank-Wolfe algorithms, a.k.a. conditional gradient algorithms, solve constrained optimization p...
Understanding intelligence and how it allows humans to learn, to make decision and form memories, is...
In this thesis, we theoretically analyze the ability of neural networks trained by gradient descent ...
A number of results have recently demonstrated the benefits of incorporating various constraints whe...
In the recent decade, deep neural networks have solved ever more complex tasks across many fronts in...
In this thesis, we focus on Frank-Wolfe (a.k.a. Conditional Gradient) algorithms, a family of iterat...
International audienceDecentralized optimization algorithms have received much attention due to the ...
Classical optimization techniques have found widespread use in machine learning. Convex optimization...
Machine learning is a technology developed for extracting predictive models from data so as to be ...
Presented on February 11, 2019 at 11:00 a.m. as part of the ARC12 Distinguished Lecture in the Klaus...
International audienceIn a series of recent theoretical works, it was shown that strongly over-param...
Aiming at convex optimization under structural constraints, this work introduces and analyzes a vari...
Large-scale machine learning problems can be reduced to non-convex optimization problems if state-of...
We revisit the Frank-Wolfe algorithm for constrained convex optimization and show that it can be imp...
Learning a deep neural network requires solving a challenging optimization problem: it is a high-dim...
The Frank-Wolfe algorithms, a.k.a. conditional gradient algorithms, solve constrained optimization p...
Understanding intelligence and how it allows humans to learn, to make decision and form memories, is...
In this thesis, we theoretically analyze the ability of neural networks trained by gradient descent ...
A number of results have recently demonstrated the benefits of incorporating various constraints whe...
In the recent decade, deep neural networks have solved ever more complex tasks across many fronts in...
In this thesis, we focus on Frank-Wolfe (a.k.a. Conditional Gradient) algorithms, a family of iterat...
International audienceDecentralized optimization algorithms have received much attention due to the ...
Classical optimization techniques have found widespread use in machine learning. Convex optimization...
Machine learning is a technology developed for extracting predictive models from data so as to be ...
Presented on February 11, 2019 at 11:00 a.m. as part of the ARC12 Distinguished Lecture in the Klaus...
International audienceIn a series of recent theoretical works, it was shown that strongly over-param...
Aiming at convex optimization under structural constraints, this work introduces and analyzes a vari...
Large-scale machine learning problems can be reduced to non-convex optimization problems if state-of...