Transfer learning has become a popular task adaptation method in the era of foundation models. However, many foundation models require large storage and computing resources, which makes off-the-shelf deployment impractical. Post-training compression techniques such as pruning and quantization can help lower deployment costs. Unfortunately, the resulting performance degradation limits the usability and benefits of such techniques. To close this performance gap, we propose CrAFT, a simple fine-tuning framework that enables effective post-training network compression. In CrAFT, users simply employ the default fine-tuning schedule along with sharpness minimization objective, simultaneously facilitating task adaptation and compression-friendline...
The success of overparameterized deep neural networks (DNNs) poses a great challenge to deploy compu...
The growing size of neural language models has led to increased attention in model compression. The ...
Multi-Task Learning (MTL) is a widely-used and powerful learning paradigm for training deep neural n...
Transfer learning has become a popular task adaptation method in the era of foundation models. Howev...
Model compression by way of parameter pruning, quantization, or distillation has recently gained pop...
End-to-end deep trainable models are about to exceed the performance of the traditional handcrafted ...
Fine-tuning BERT-based models is resource-intensive in memory, computation, and time. While many pri...
We present RECLIP (Resource-efficient CLIP), a simple method that minimizes computational resource f...
177 pagesThe field of computer vision has benefited tremendously from an unusual blessing: a baselin...
As Deep Neural Networks (DNNs) usually are overparameterized and have millions of weight parameters,...
Existing fine-tuning methods either tune all parameters of the pre-trained model (full fine-tuning),...
There are growing interests in adapting large-scale language models using parameter-efficient fine-t...
We propose a novel method for training a neural network for image classification to reduce input dat...
As a dominant paradigm, fine-tuning a pre-trained model on the target data is widely used in many de...
The current modus operandi in adapting pre-trained models involves updating all the backbone paramet...
The success of overparameterized deep neural networks (DNNs) poses a great challenge to deploy compu...
The growing size of neural language models has led to increased attention in model compression. The ...
Multi-Task Learning (MTL) is a widely-used and powerful learning paradigm for training deep neural n...
Transfer learning has become a popular task adaptation method in the era of foundation models. Howev...
Model compression by way of parameter pruning, quantization, or distillation has recently gained pop...
End-to-end deep trainable models are about to exceed the performance of the traditional handcrafted ...
Fine-tuning BERT-based models is resource-intensive in memory, computation, and time. While many pri...
We present RECLIP (Resource-efficient CLIP), a simple method that minimizes computational resource f...
177 pagesThe field of computer vision has benefited tremendously from an unusual blessing: a baselin...
As Deep Neural Networks (DNNs) usually are overparameterized and have millions of weight parameters,...
Existing fine-tuning methods either tune all parameters of the pre-trained model (full fine-tuning),...
There are growing interests in adapting large-scale language models using parameter-efficient fine-t...
We propose a novel method for training a neural network for image classification to reduce input dat...
As a dominant paradigm, fine-tuning a pre-trained model on the target data is widely used in many de...
The current modus operandi in adapting pre-trained models involves updating all the backbone paramet...
The success of overparameterized deep neural networks (DNNs) poses a great challenge to deploy compu...
The growing size of neural language models has led to increased attention in model compression. The ...
Multi-Task Learning (MTL) is a widely-used and powerful learning paradigm for training deep neural n...