Rapid advances in data collection and processing capabilities have allowed for the use of increasingly complex models that give rise to nonconvex optimization problems. These formulations, however, can be arbitrarily difficult to solve in general, in the sense that even simply verifying that a given point is a local minimum can be NP-hard [1]. Still, some relatively simple algorithms have been shown to lead to surprisingly good empirical results in many contexts of interest. Perhaps the most prominent example is the success of the backpropagation algorithm for training neural networks. Several recent works have pursued rigorous analytical justification for this phenomenon by studying the structure of the nonconvex optimization problems and ...
Recently, we proposed to transform the outputs of each hidden neuron in a multi-layer perceptron net...
Optimization is the key component of deep learning. Increasing depth, which is vital for reaching a...
© 2019 Massachusetts Institute of Technology. For nonconvex optimization in machine learning, this a...
Federated learning is a useful framework for centralized learning from distributed data under practi...
In recent centralized nonconvex distributed learning and federated learning, local methods are one o...
Solving large scale optimization problems, such as neural networks training, can present many challe...
Abstract. Recently, we proposed to transform the outputs of each hidden neu-ron in a multi-layer per...
In the recent decade, deep neural networks have solved ever more complex tasks across many fronts in...
Presented on February 11, 2019 at 11:00 a.m. as part of the ARC12 Distinguished Lecture in the Klaus...
Machine learning and reinforcement learning have achieved tremendous success in solving problems in ...
Decentralized optimization algorithms have attracted intensive interests recently, as it has a balan...
In order to be provably convergent towards a second-order stationary point, optimization methods app...
We consider the fundamental problem in nonconvex optimization of efficiently reaching a stationary p...
First-order methods are gaining substantial interest in the past two decades because of their superi...
In this paper, we provide new results and algorithms (including backtracking versions of Nesterov ac...
Recently, we proposed to transform the outputs of each hidden neuron in a multi-layer perceptron net...
Optimization is the key component of deep learning. Increasing depth, which is vital for reaching a...
© 2019 Massachusetts Institute of Technology. For nonconvex optimization in machine learning, this a...
Federated learning is a useful framework for centralized learning from distributed data under practi...
In recent centralized nonconvex distributed learning and federated learning, local methods are one o...
Solving large scale optimization problems, such as neural networks training, can present many challe...
Abstract. Recently, we proposed to transform the outputs of each hidden neu-ron in a multi-layer per...
In the recent decade, deep neural networks have solved ever more complex tasks across many fronts in...
Presented on February 11, 2019 at 11:00 a.m. as part of the ARC12 Distinguished Lecture in the Klaus...
Machine learning and reinforcement learning have achieved tremendous success in solving problems in ...
Decentralized optimization algorithms have attracted intensive interests recently, as it has a balan...
In order to be provably convergent towards a second-order stationary point, optimization methods app...
We consider the fundamental problem in nonconvex optimization of efficiently reaching a stationary p...
First-order methods are gaining substantial interest in the past two decades because of their superi...
In this paper, we provide new results and algorithms (including backtracking versions of Nesterov ac...
Recently, we proposed to transform the outputs of each hidden neuron in a multi-layer perceptron net...
Optimization is the key component of deep learning. Increasing depth, which is vital for reaching a...
© 2019 Massachusetts Institute of Technology. For nonconvex optimization in machine learning, this a...