Pre-trained language models (PLMs) have demonstrated impressive performance across various downstream NLP tasks. Nevertheless, the resource requirements of pre-training large language models in terms of memory and training compute pose significant challenges. Furthermore, due to the substantial resources required, many PLM weights are confidential. Consequently, users are compelled to share their data with model owners for fine-tuning on specific tasks. To overcome the limitations, we introduce Plug-in External Memory Adaptation (PEMA), a Parameter-Efficient Fine-Tuning (PEFT) approach designed for fine-tuning PLMs without the need for all weights. PEMA can be integrated into the context representation of test data during inference to execu...
Widely used language models (LMs) are typically built by scaling up a two-stage training pipeline: a...
Fine-tuning BERT-based models is resource-intensive in memory, computation, and time. While many pri...
Parameter-shared pre-trained language models (PLMs) have emerged as a successful approach in resourc...
Adapting pretrained language models to novel domains, such as clinical applications, traditionally i...
Pretrained large language models (LLMs) are strong in-context learners that are able to perform few-...
Prior work shows that it is possible to expand pretrained Masked Language Models (MLMs) to new langu...
There are growing interests in adapting large-scale language models using parameter-efficient fine-t...
Recent advances in NLP are brought by a range of large-scale pretrained language models (PLMs). Thes...
Adopting a two-stage paradigm of pretraining followed by fine-tuning, Pretrained Language Models (PL...
Real-world business applications require a trade-off between language model performance and size. We...
Recent advancements in Large Language Models (LLMs) have enabled the development of a single model c...
The reusability of state-of-the-art Pre-trained Language Models (PLMs) is often limited by their gen...
Open-sourced large language models (LLMs) have demonstrated remarkable efficacy in various tasks wit...
Fine-tuning large language models for different tasks can be costly and inefficient, and even method...
Large language models (LLMs) and vision language models (VLMs) demonstrate excellent performance on ...
Widely used language models (LMs) are typically built by scaling up a two-stage training pipeline: a...
Fine-tuning BERT-based models is resource-intensive in memory, computation, and time. While many pri...
Parameter-shared pre-trained language models (PLMs) have emerged as a successful approach in resourc...
Adapting pretrained language models to novel domains, such as clinical applications, traditionally i...
Pretrained large language models (LLMs) are strong in-context learners that are able to perform few-...
Prior work shows that it is possible to expand pretrained Masked Language Models (MLMs) to new langu...
There are growing interests in adapting large-scale language models using parameter-efficient fine-t...
Recent advances in NLP are brought by a range of large-scale pretrained language models (PLMs). Thes...
Adopting a two-stage paradigm of pretraining followed by fine-tuning, Pretrained Language Models (PL...
Real-world business applications require a trade-off between language model performance and size. We...
Recent advancements in Large Language Models (LLMs) have enabled the development of a single model c...
The reusability of state-of-the-art Pre-trained Language Models (PLMs) is often limited by their gen...
Open-sourced large language models (LLMs) have demonstrated remarkable efficacy in various tasks wit...
Fine-tuning large language models for different tasks can be costly and inefficient, and even method...
Large language models (LLMs) and vision language models (VLMs) demonstrate excellent performance on ...
Widely used language models (LMs) are typically built by scaling up a two-stage training pipeline: a...
Fine-tuning BERT-based models is resource-intensive in memory, computation, and time. While many pri...
Parameter-shared pre-trained language models (PLMs) have emerged as a successful approach in resourc...