In many high-dimensional learning problems, only some parts of an observation are important to the prediction task; for example, the cues to correctly categorizing a document may lie in a handful of its sentences. We introduce a learning algorithm that exploits this intuition by encod-ing it in a regularizer. Specifically, we apply the sparse overlapping group lasso with one group for every bundle of features occurring together in a training-data sentence, leading to thousands to millions of overlapping groups. We show how to efficiently solve the resulting optimization chal-lenge using the alternating directions method of multipliers. We find that the resulting method significantly outperforms competitive baselines (standard ridge, lasso, ...
International audienceRegression with group-sparsity penalty plays a central role in high-dimensiona...
In many real-world applications of supervised learning, only a limited number of labeled examples ar...
Data augmentation with Mixup (Zhang et al. 2018) has shown to be an effective model regularizer for ...
In many high-dimensional learning problems, only some parts of an observation are important to the p...
We introduce three linguistically moti-vated structured regularizers based on parse trees, topics, a...
We consider the problem of learning a sparse multi-task regression, where the structure in the outpu...
We consider the problem of learning a sparse multi-task regression, where the structure in the outpu...
The paper considers supervised learning problems of labeled data with grouped input features. The gr...
l2,1-norm is an effective regularization to enforce a simple group sparsity for feature learning. To...
This paper investigates, from information theoretic grounds, a learning problem based on the princip...
We present a data dependent generalization bound for a large class of regularized algorithms which i...
Group LASSO is widely used to enforce the structural sparsity, which achieves the sparsity at the in...
Binary logistic regression with a sparsity constraint on the solution plays a vital role in many hig...
We develop a dictionary learning method which is (i) online, (ii) enables overlapping group structur...
13 pages, 7 figures. Submitted for publicationThis paper investigates, from information theoretic gr...
International audienceRegression with group-sparsity penalty plays a central role in high-dimensiona...
In many real-world applications of supervised learning, only a limited number of labeled examples ar...
Data augmentation with Mixup (Zhang et al. 2018) has shown to be an effective model regularizer for ...
In many high-dimensional learning problems, only some parts of an observation are important to the p...
We introduce three linguistically moti-vated structured regularizers based on parse trees, topics, a...
We consider the problem of learning a sparse multi-task regression, where the structure in the outpu...
We consider the problem of learning a sparse multi-task regression, where the structure in the outpu...
The paper considers supervised learning problems of labeled data with grouped input features. The gr...
l2,1-norm is an effective regularization to enforce a simple group sparsity for feature learning. To...
This paper investigates, from information theoretic grounds, a learning problem based on the princip...
We present a data dependent generalization bound for a large class of regularized algorithms which i...
Group LASSO is widely used to enforce the structural sparsity, which achieves the sparsity at the in...
Binary logistic regression with a sparsity constraint on the solution plays a vital role in many hig...
We develop a dictionary learning method which is (i) online, (ii) enables overlapping group structur...
13 pages, 7 figures. Submitted for publicationThis paper investigates, from information theoretic gr...
International audienceRegression with group-sparsity penalty plays a central role in high-dimensiona...
In many real-world applications of supervised learning, only a limited number of labeled examples ar...
Data augmentation with Mixup (Zhang et al. 2018) has shown to be an effective model regularizer for ...