First-order methods are gaining substantial interest in the past two decades because of their superior performance in solving today's large-scale problems. In this thesis, we study some widely used first-order methods for problems that satisfy certain structures. Specifically, in the first part, we contribute to coordinate optimization and show that greedy coordinate descent (GCD) has an implicit screening ability that usually selects coordinates that are nonzero at the solution, which explains why GCD works exceptionally well for problems that admit sparse solutions. We also extend the elegant safe-screening rule that depends on duality gap to atomic-norm regularized problems. In the second part, we study online mirror descent (OMD) with u...
In this paper we prove a new complexity bound for a variant of Accelerated Coordinate Descent Method...
Optimization is an important discipline of applied mathematics with far-reaching applications. Optim...
Large-scale machine learning problems can be reduced to non-convex optimization problems if state-of...
Large-scale optimization problems appear quite frequently in data science and machine learning appli...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
We study a family of first-order methods with momentum based on mirror descent for online convex opt...
Introduction. Recent work has shown many connections between conditional gradient and other first-or...
International audienceWe introduce and analyze a new family of first-order optimization algorithms w...
This work looks at large-scale machine learning, with a particular focus on greedy methods. A recent...
Thesis: Ph. D. in Mathematics and Operations Research, Massachusetts Institute of Technology, Depart...
We develop a modified online mirror descent framework that is suitable for building adaptive and par...
Coordinate descent with random coordinate selection is the current state of the art for many large s...
Classical global convergence results for first-order methods rely on uniform smoothness and the \L{}...
We present a new method for regularized convex optimization and analyze it under both online and sto...
We present a simple unified analysis of adaptive Mirror Descent (MD) and Follow- the-Regularized-Lea...
In this paper we prove a new complexity bound for a variant of Accelerated Coordinate Descent Method...
Optimization is an important discipline of applied mathematics with far-reaching applications. Optim...
Large-scale machine learning problems can be reduced to non-convex optimization problems if state-of...
Large-scale optimization problems appear quite frequently in data science and machine learning appli...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
We study a family of first-order methods with momentum based on mirror descent for online convex opt...
Introduction. Recent work has shown many connections between conditional gradient and other first-or...
International audienceWe introduce and analyze a new family of first-order optimization algorithms w...
This work looks at large-scale machine learning, with a particular focus on greedy methods. A recent...
Thesis: Ph. D. in Mathematics and Operations Research, Massachusetts Institute of Technology, Depart...
We develop a modified online mirror descent framework that is suitable for building adaptive and par...
Coordinate descent with random coordinate selection is the current state of the art for many large s...
Classical global convergence results for first-order methods rely on uniform smoothness and the \L{}...
We present a new method for regularized convex optimization and analyze it under both online and sto...
We present a simple unified analysis of adaptive Mirror Descent (MD) and Follow- the-Regularized-Lea...
In this paper we prove a new complexity bound for a variant of Accelerated Coordinate Descent Method...
Optimization is an important discipline of applied mathematics with far-reaching applications. Optim...
Large-scale machine learning problems can be reduced to non-convex optimization problems if state-of...