Pretrained Transformers achieve state-of-the-art performance in various code-processing tasks but may be too large to be deployed. As software development tools often incorporate modules for various purposes which may potentially use a single instance of the pretrained model, it appears relevant to utilize parameter-efficient fine-tuning for the pretrained models of code. In this work, we test two widely used approaches, adapters and LoRA, which were initially tested on NLP tasks, on four code-processing tasks. We find that though the efficient fine-tuning approaches may achieve comparable or higher performance than the standard, full, fine-tuning in code understanding tasks, they underperform full fine-tuning in code-generative tasks. Thes...
Fine-tuning BERT-based models is resource-intensive in memory, computation, and time. While many pri...
Gigantic pre-trained models have become central to natural language processing (NLP), serving as the...
The field of natural language processing (NLP) has recently undergone a paradigm shift. Since the in...
We introduce BitFit, a sparse-finetuning method where only the bias-terms of the model (or a subset ...
Large Transformer models achieved the state-of-the-art status for Natural Language Understanding tas...
Transformers are responsible for the vast majority of recent advances in natural language processing...
The current modus operandi in adapting pre-trained models involves updating all the backbone paramet...
Transformer-based pre-trained models with millions of parameters require large storage. Recent appro...
Pre-trained transformers have rapidly become very popular in the Natural Language Processing (NLP) c...
Pre-trained language models received extensive attention in recent years. However, it is still chall...
Parameter-efficient fine-tuning methods (PEFTs) offer the promise of adapting large pre-trained mode...
Transformer models cannot easily scale to long sequences due to their O(N^2) time and space complexi...
We present Generalized LoRA (GLoRA), an advanced approach for universal parameter-efficient fine-tun...
Transformer based models are used to achieve state-of-the-art performance on various deep learning t...
State-of-the-art pretrained NLP models contain a hundred million to trillion parameters. Adapters pr...
Fine-tuning BERT-based models is resource-intensive in memory, computation, and time. While many pri...
Gigantic pre-trained models have become central to natural language processing (NLP), serving as the...
The field of natural language processing (NLP) has recently undergone a paradigm shift. Since the in...
We introduce BitFit, a sparse-finetuning method where only the bias-terms of the model (or a subset ...
Large Transformer models achieved the state-of-the-art status for Natural Language Understanding tas...
Transformers are responsible for the vast majority of recent advances in natural language processing...
The current modus operandi in adapting pre-trained models involves updating all the backbone paramet...
Transformer-based pre-trained models with millions of parameters require large storage. Recent appro...
Pre-trained transformers have rapidly become very popular in the Natural Language Processing (NLP) c...
Pre-trained language models received extensive attention in recent years. However, it is still chall...
Parameter-efficient fine-tuning methods (PEFTs) offer the promise of adapting large pre-trained mode...
Transformer models cannot easily scale to long sequences due to their O(N^2) time and space complexi...
We present Generalized LoRA (GLoRA), an advanced approach for universal parameter-efficient fine-tun...
Transformer based models are used to achieve state-of-the-art performance on various deep learning t...
State-of-the-art pretrained NLP models contain a hundred million to trillion parameters. Adapters pr...
Fine-tuning BERT-based models is resource-intensive in memory, computation, and time. While many pri...
Gigantic pre-trained models have become central to natural language processing (NLP), serving as the...
The field of natural language processing (NLP) has recently undergone a paradigm shift. Since the in...