Model sparsification in deep learning promotes simpler, more interpretable models with fewer parameters. This not only reduces the model's memory footprint and computational needs but also shortens inference time. This work focuses on creating sparse models optimized for multiple tasks with fewer parameters. These parsimonious models also possess the potential to match or outperform dense models in terms of performance. In this work, we introduce channel-wise l1/l2 group sparsity in the shared convolutional layers parameters (or weights) of the multi-task learning model. This approach facilitates the removal of extraneous groups i.e., channels (due to l1 regularization) and also imposes a penalty on the weights, further enhancing the learni...
Sparsity plays a key role in machine learning for several reasons including interpretability. Interp...
International audience—Spatially-sparse predictors are good models for brain decoding: they give acc...
Structured sparse learning has become a popular and mature research field. Among all structured spar...
Sparsity has become one of the promising methods to compress and accelerate Deep Neural Networks (DN...
With the dramatically increased number of parameters in language models, sparsity methods have recei...
The growing energy and performance costs of deep learning have driven the community to reduce the si...
Overparameterized neural networks generalize well but are expensive to train. Ideally, one would lik...
Deep learning has been empirically successful in recent years thanks to the extremely over-parameter...
The training of sparse neural networks is becoming an increasingly important tool for reducing the ...
The pre-training and fine-tuning paradigm has contributed to a number of breakthroughs in Natural La...
We identify and prove a general principle: $L_1$ sparsity can be achieved using a redundant parametr...
The articles in this special section focus on learning adaptive models. Over the past few years, spa...
In this paper, we address the challenging task of simultaneously optimizing (i) the weights of a neu...
Gigantic pre-trained models have become central to natural language processing (NLP), serving as the...
In this paper, we address the challenging task of simultaneously optimizing (i) the weights of a neu...
Sparsity plays a key role in machine learning for several reasons including interpretability. Interp...
International audience—Spatially-sparse predictors are good models for brain decoding: they give acc...
Structured sparse learning has become a popular and mature research field. Among all structured spar...
Sparsity has become one of the promising methods to compress and accelerate Deep Neural Networks (DN...
With the dramatically increased number of parameters in language models, sparsity methods have recei...
The growing energy and performance costs of deep learning have driven the community to reduce the si...
Overparameterized neural networks generalize well but are expensive to train. Ideally, one would lik...
Deep learning has been empirically successful in recent years thanks to the extremely over-parameter...
The training of sparse neural networks is becoming an increasingly important tool for reducing the ...
The pre-training and fine-tuning paradigm has contributed to a number of breakthroughs in Natural La...
We identify and prove a general principle: $L_1$ sparsity can be achieved using a redundant parametr...
The articles in this special section focus on learning adaptive models. Over the past few years, spa...
In this paper, we address the challenging task of simultaneously optimizing (i) the weights of a neu...
Gigantic pre-trained models have become central to natural language processing (NLP), serving as the...
In this paper, we address the challenging task of simultaneously optimizing (i) the weights of a neu...
Sparsity plays a key role in machine learning for several reasons including interpretability. Interp...
International audience—Spatially-sparse predictors are good models for brain decoding: they give acc...
Structured sparse learning has become a popular and mature research field. Among all structured spar...