Large, pre-trained models are problematic to use in resource constrained applications. Fortunately, task-aware structured pruning methods offer a solution. These approaches reduce model size by dropping structural units like layers and attention heads in a manner that takes into account the end-task. However, these pruning algorithms require more task-specific data than is typically available. We propose a framework which combines structured pruning with transfer learning to reduce the need for task-specific data. Our empirical results answer questions such as: How should the two tasks be coupled? What parameters should be transferred? And, when during training should transfer learning be introduced? Leveraging these insights, we demonstrat...
Inductive transfer learning aims to learn from a small amount of training data for the target task b...
Model sparsification in deep learning promotes simpler, more interpretable models with fewer paramet...
Network pruning is an important research field aiming at reducing computational costs of neural netw...
The growing size of neural language models has led to increased attention in model compression. The ...
The increasing of pre-trained models has significantly facilitated the performance on limited data t...
Pre-trained Language Models (PLMs) have achieved great success in various Natural Language Processin...
Large-scale pre-trained models have been remarkably successful in resolving downstream tasks. Noneth...
Sparsity has become one of the promising methods to compress and accelerate Deep Neural Networks (DN...
Structured pruning is a popular method to reduce the cost of convolutional neural networks, that are...
Pruning is a compression method which aims to improve the efficiency of neural networks by reducing ...
Transfer-learning methods aim to improve performance in a data-scarce target domain using a model pr...
International audienceThe wide deployment of Machine Learning models is an essential evolution of Ar...
Introducing sparsity in a neural network has been an efficient way to reduce its complexity while ke...
Neural models have seen great success in computer vision in recent years, especially in fundamental...
Transfer learning, where a model is first pre-trained on a data-rich task before being fine-tuned on...
Inductive transfer learning aims to learn from a small amount of training data for the target task b...
Model sparsification in deep learning promotes simpler, more interpretable models with fewer paramet...
Network pruning is an important research field aiming at reducing computational costs of neural netw...
The growing size of neural language models has led to increased attention in model compression. The ...
The increasing of pre-trained models has significantly facilitated the performance on limited data t...
Pre-trained Language Models (PLMs) have achieved great success in various Natural Language Processin...
Large-scale pre-trained models have been remarkably successful in resolving downstream tasks. Noneth...
Sparsity has become one of the promising methods to compress and accelerate Deep Neural Networks (DN...
Structured pruning is a popular method to reduce the cost of convolutional neural networks, that are...
Pruning is a compression method which aims to improve the efficiency of neural networks by reducing ...
Transfer-learning methods aim to improve performance in a data-scarce target domain using a model pr...
International audienceThe wide deployment of Machine Learning models is an essential evolution of Ar...
Introducing sparsity in a neural network has been an efficient way to reduce its complexity while ke...
Neural models have seen great success in computer vision in recent years, especially in fundamental...
Transfer learning, where a model is first pre-trained on a data-rich task before being fine-tuned on...
Inductive transfer learning aims to learn from a small amount of training data for the target task b...
Model sparsification in deep learning promotes simpler, more interpretable models with fewer paramet...
Network pruning is an important research field aiming at reducing computational costs of neural netw...