We propose a computational framework named iterative local adaptive majorize-minimization (I-LAMM) to simultaneously control algorithmic complexity and statistical error when fitting high dimensional models. I-LAMM is a two-stage algorithmic implementation of the local linear approximation to a family of folded concave penalized quasi-likelihood. The first stage solves a convex program with a crude precision tolerance to obtain a coarse initial estimator, which is further refined in the second stage by iteratively solving a sequence of convex programs with smaller precision tolerances. Theoretically, we establish a phase transition: the first stage has a sublinear iteration complexity, while the second stage achieves an improved linear rate...
Folded concave penalization methods have been shown to enjoy the strong oracle property for high-dim...
Many important machine learning problems are modeled and solved via semidefinite programs; examples ...
Sparse learning models typically combine a smooth loss with a nonsmooth penalty, such as trace norm....
The era of machine learning features large datasets that have high dimension of features. This leads...
The dissertation addresses the research topics of machine learning outlined below. We developed the ...
Nowadays, the explosive data scale increase provides an unprecedented opportunity to apply machine l...
The file attached to this record is the author's final peer reviewed version. The Publisher's final ...
Abstract. Majorization-minimization algorithms consist of successively minimizing a sequence of uppe...
International audienceMajorization-minimization algorithms consist of successively minimizing a sequ...
This thesis shows how we can exploit low-dimensional structure in high-dimensional statistics and ma...
Modern machine learning has made significant breakthroughs for scientific and technological applicat...
Modern technological advances have prompted massive scale data collection in manymodern fields such ...
We consider machine learning techniques to develop low-latency approximate solutions for a class of ...
Many important machine learning problems are modeled and solved via semidefinite programs; examples ...
Thesis: Ph. D. in Mathematics and Operations Research, Massachusetts Institute of Technology, Depart...
Folded concave penalization methods have been shown to enjoy the strong oracle property for high-dim...
Many important machine learning problems are modeled and solved via semidefinite programs; examples ...
Sparse learning models typically combine a smooth loss with a nonsmooth penalty, such as trace norm....
The era of machine learning features large datasets that have high dimension of features. This leads...
The dissertation addresses the research topics of machine learning outlined below. We developed the ...
Nowadays, the explosive data scale increase provides an unprecedented opportunity to apply machine l...
The file attached to this record is the author's final peer reviewed version. The Publisher's final ...
Abstract. Majorization-minimization algorithms consist of successively minimizing a sequence of uppe...
International audienceMajorization-minimization algorithms consist of successively minimizing a sequ...
This thesis shows how we can exploit low-dimensional structure in high-dimensional statistics and ma...
Modern machine learning has made significant breakthroughs for scientific and technological applicat...
Modern technological advances have prompted massive scale data collection in manymodern fields such ...
We consider machine learning techniques to develop low-latency approximate solutions for a class of ...
Many important machine learning problems are modeled and solved via semidefinite programs; examples ...
Thesis: Ph. D. in Mathematics and Operations Research, Massachusetts Institute of Technology, Depart...
Folded concave penalization methods have been shown to enjoy the strong oracle property for high-dim...
Many important machine learning problems are modeled and solved via semidefinite programs; examples ...
Sparse learning models typically combine a smooth loss with a nonsmooth penalty, such as trace norm....