With the increasing prevalence of Large Language Models, traditional full fine-tuning approaches face growing challenges, especially in memory-intensive tasks. This paper investigates the potential of Parameter-Efficient Fine-Tuning, focusing on Low-Rank Adaptation (LoRA), for complex and under-explored multilingual summarisation tasks. We conduct an extensive study across different data availability scenarios, including full-data, low-data, and cross-lingual transfer, leveraging models of different sizes. Our findings reveal that LoRA lags behind full fine-tuning when trained with full data, however, it excels in low-data scenarios and cross-lingual transfer. Interestingly, as models scale up, the performance gap between LoRA and full fine...
Multilingual pre-trained language models (PLMs) have demonstrated impressive performance on several ...
In this paper, we present Delta-LoRA, which is a novel parameter-efficient approach to fine-tune lar...
Pre-trained multilingual models, such as mBERT, XLM-R and mT5, are used to improve the performance o...
Fine-tuning the entire set of parameters of a large pretrained model has become the mainstream appro...
Pre-trained multilingual language models show significant performance gains for zero-shot cross-ling...
Large Language Models (LLMs), trained predominantly on extensive English data, often exhibit limitat...
We present Generalized LoRA (GLoRA), an advanced approach for universal parameter-efficient fine-tun...
Large language models (LLMs) and vision language models (VLMs) demonstrate excellent performance on ...
Pre-trained language models (PLMs) have demonstrated impressive performance across various downstrea...
Multilingual language models are widely used to extend NLP systems to low-resource languages. Howeve...
In this paper, we explore the challenging problem of performing a generative task (i.e., summarizati...
Pretrained multilingual language models have become a common tool in transferring NLP capabilities t...
Cross-lingual transfer learning with large multilingual pre-trained models can be an effective appro...
State-of-the-art neural (re)rankers are notoriously data hungry which - given the lack of large-scal...
Adapting pretrained language models to novel domains, such as clinical applications, traditionally i...
Multilingual pre-trained language models (PLMs) have demonstrated impressive performance on several ...
In this paper, we present Delta-LoRA, which is a novel parameter-efficient approach to fine-tune lar...
Pre-trained multilingual models, such as mBERT, XLM-R and mT5, are used to improve the performance o...
Fine-tuning the entire set of parameters of a large pretrained model has become the mainstream appro...
Pre-trained multilingual language models show significant performance gains for zero-shot cross-ling...
Large Language Models (LLMs), trained predominantly on extensive English data, often exhibit limitat...
We present Generalized LoRA (GLoRA), an advanced approach for universal parameter-efficient fine-tun...
Large language models (LLMs) and vision language models (VLMs) demonstrate excellent performance on ...
Pre-trained language models (PLMs) have demonstrated impressive performance across various downstrea...
Multilingual language models are widely used to extend NLP systems to low-resource languages. Howeve...
In this paper, we explore the challenging problem of performing a generative task (i.e., summarizati...
Pretrained multilingual language models have become a common tool in transferring NLP capabilities t...
Cross-lingual transfer learning with large multilingual pre-trained models can be an effective appro...
State-of-the-art neural (re)rankers are notoriously data hungry which - given the lack of large-scal...
Adapting pretrained language models to novel domains, such as clinical applications, traditionally i...
Multilingual pre-trained language models (PLMs) have demonstrated impressive performance on several ...
In this paper, we present Delta-LoRA, which is a novel parameter-efficient approach to fine-tune lar...
Pre-trained multilingual models, such as mBERT, XLM-R and mT5, are used to improve the performance o...