Parameter-efficient tuning (PET) has been widely explored in recent years because it tunes much fewer parameters (PET modules) than full-parameter fine-tuning (FT) while still stimulating sufficient knowledge from large language models (LLMs) for downstream tasks. Moreover, when PET is employed to serve multiple tasks, different task-specific PET modules can be built on a frozen LLM, avoiding redundant LLM deployments. Although PET significantly reduces the cost of tuning and deploying LLMs, its inference still suffers from the computational bottleneck of LLMs. To address the above issue, we propose an effective PET framework based on compressed LLMs, named "CPET". In CPET, we evaluate the impact of mainstream LLM compression techniques on ...
We consider the problem of accurate sparse fine-tuning of large language models (LLMs), that is, fin...
While GPTs with traditional fine-tuning fail to achieve strong results on natural language understan...
Pre-trained Language Models (PLMs) have achieved great success in various Natural Language Processin...
Continual learning necessitates the continual adaptation of models to newly emerging tasks while min...
Fine-tuning BERT-based models is resource-intensive in memory, computation, and time. While many pri...
Adapting large pre-trained models (PTMs) through fine-tuning imposes prohibitive computational and s...
There are growing interests in adapting large-scale language models using parameter-efficient fine-t...
Large language models (LLMs), while transformative for NLP, come with significant computational dema...
Pre-trained language models (PLMs) have demonstrated impressive performance across various downstrea...
Pretrained large language models (LLMs) are strong in-context learners that are able to perform few-...
When scaled to hundreds of billions of parameters, pretrained language models such as GPT-3 (Brown e...
Recent advancements in Large Language Models (LLMs) have enabled the development of a single model c...
Parameter-shared pre-trained language models (PLMs) have emerged as a successful approach in resourc...
Advancements in deep learning are often associated with increasing model sizes. The model size drama...
Multilingual models are often particularly dependent on scaling to generalize to a growing number of...
We consider the problem of accurate sparse fine-tuning of large language models (LLMs), that is, fin...
While GPTs with traditional fine-tuning fail to achieve strong results on natural language understan...
Pre-trained Language Models (PLMs) have achieved great success in various Natural Language Processin...
Continual learning necessitates the continual adaptation of models to newly emerging tasks while min...
Fine-tuning BERT-based models is resource-intensive in memory, computation, and time. While many pri...
Adapting large pre-trained models (PTMs) through fine-tuning imposes prohibitive computational and s...
There are growing interests in adapting large-scale language models using parameter-efficient fine-t...
Large language models (LLMs), while transformative for NLP, come with significant computational dema...
Pre-trained language models (PLMs) have demonstrated impressive performance across various downstrea...
Pretrained large language models (LLMs) are strong in-context learners that are able to perform few-...
When scaled to hundreds of billions of parameters, pretrained language models such as GPT-3 (Brown e...
Recent advancements in Large Language Models (LLMs) have enabled the development of a single model c...
Parameter-shared pre-trained language models (PLMs) have emerged as a successful approach in resourc...
Advancements in deep learning are often associated with increasing model sizes. The model size drama...
Multilingual models are often particularly dependent on scaling to generalize to a growing number of...
We consider the problem of accurate sparse fine-tuning of large language models (LLMs), that is, fin...
While GPTs with traditional fine-tuning fail to achieve strong results on natural language understan...
Pre-trained Language Models (PLMs) have achieved great success in various Natural Language Processin...