We study stochastic optimization problems when the data is sparse, which is in a sense dual to the current understanding of high-dimensional statistical learning and optimization. We highlight both the difficulties—in terms of increased sample complexity that sparse data necessitates—and the potential benefits, in terms of allowing parallelism and asynchrony in the design of algorithms. Concretely, we derive matching upper and lower bounds on the minimax rate for optimization and learning with sparse data, and we exhibit algorithms achieving these rates. We also show how leveraging sparsity leads to (still minimax optimal) parallel and asynchronous algorithms, providing experimental evidence complementing our theoretical results on several ...
Nowadays, the explosive data scale increase provides an unprecedented opportunity to apply machine l...
Many statistical learning problems can be posed as minimization of a sum of two convex functions, on...
Appears in Advances in Neural Information Processing Systems 30 (NIPS 2017), 28 pagesInternational a...
We study stochastic optimization problems when the data is sparse, which is in a sense dual to curre...
In this paper, we focus on Stochastic Composite Optimization (SCO) for sparse learning that aims to ...
<p> In this paper, we study Stochastic Composite Optimization (SCO) for sparse learning that aims t...
Thesis: S.M., Massachusetts Institute of Technology, Sloan School of Management, Operations Research...
Classically, the performance of estimators in statistical learning problems is measured in terms of ...
We formulate the sparse classification problem of n samples with p features as a binary convex optim...
Sparse estimation methods are aimed at using or obtaining parsimonious representations of data or mo...
This thesis proposes parallel and distributed algorithms for solving very largescale sparse optimiza...
Modern machine learning systems pose several new statistical, scalability, privacy and ethical chall...
Modern technological advances have prompted massive scale data collection in manymodern fields such ...
Classical optimization techniques have found widespread use in machine learning. Convex optimization...
International audienceIn distributed optimization for large-scale learning, a major performance limi...
Nowadays, the explosive data scale increase provides an unprecedented opportunity to apply machine l...
Many statistical learning problems can be posed as minimization of a sum of two convex functions, on...
Appears in Advances in Neural Information Processing Systems 30 (NIPS 2017), 28 pagesInternational a...
We study stochastic optimization problems when the data is sparse, which is in a sense dual to curre...
In this paper, we focus on Stochastic Composite Optimization (SCO) for sparse learning that aims to ...
<p> In this paper, we study Stochastic Composite Optimization (SCO) for sparse learning that aims t...
Thesis: S.M., Massachusetts Institute of Technology, Sloan School of Management, Operations Research...
Classically, the performance of estimators in statistical learning problems is measured in terms of ...
We formulate the sparse classification problem of n samples with p features as a binary convex optim...
Sparse estimation methods are aimed at using or obtaining parsimonious representations of data or mo...
This thesis proposes parallel and distributed algorithms for solving very largescale sparse optimiza...
Modern machine learning systems pose several new statistical, scalability, privacy and ethical chall...
Modern technological advances have prompted massive scale data collection in manymodern fields such ...
Classical optimization techniques have found widespread use in machine learning. Convex optimization...
International audienceIn distributed optimization for large-scale learning, a major performance limi...
Nowadays, the explosive data scale increase provides an unprecedented opportunity to apply machine l...
Many statistical learning problems can be posed as minimization of a sum of two convex functions, on...
Appears in Advances in Neural Information Processing Systems 30 (NIPS 2017), 28 pagesInternational a...