Pre-trained Language Models are widely used in many important real-world applications. However, recent studies show that these models can encode social biases from large pre-training corpora and even amplify biases in downstream applications. To address this challenge, we propose Co$^2$PT, an efficient and effective debias-while-prompt tuning method for mitigating biases via counterfactual contrastive prompt tuning on downstream tasks. Our experiments conducted on three extrinsic bias benchmarks demonstrate the effectiveness of Co$^2$PT on bias mitigation during the prompt tuning process and its adaptability to existing upstream debiased language models. These findings indicate the strength of Co$^2$PT and provide promising avenues for furt...
Recent works have shown promising results of prompt tuning in stimulating pre-trained language model...
Warning: this paper contains model outputs exhibiting offensiveness and biases. Recently pre-trained...
When trained on large, unfiltered crawls from the Internet, language models pick up and reproduce al...
As the representation capability of Pre-trained Language Models (PLMs) improve, there is growing con...
Despite their impressive performance in a wide range of NLP tasks, Large Language Models (LLMs) have...
Pre-trained language models reflect the inherent social biases of their training corpus. Many method...
Cheap-to-Build Very Large-Language Models (CtB-LLMs) with affordable training are emerging as the ne...
Pre-training a language model and then fine-tuning it for downstream tasks has demonstrated state-of...
Large pre-trained language models are successfully being used in a variety of tasks, across many lan...
Language model fine-tuning is essential for modern natural language processing, but is computational...
Pretrained language models can be effectively stimulated by textual prompts or demonstrations, espec...
Group bias in natural language processing tasks manifests as disparities in system error rates acros...
We introduce BitFit, a sparse-finetuning method where only the bias-terms of the model (or a subset ...
Language models (LMs) exhibit and amplify many types of undesirable biases learned from the training...
Several prior studies have suggested that word frequency biases can cause the Bert model to learn in...
Recent works have shown promising results of prompt tuning in stimulating pre-trained language model...
Warning: this paper contains model outputs exhibiting offensiveness and biases. Recently pre-trained...
When trained on large, unfiltered crawls from the Internet, language models pick up and reproduce al...
As the representation capability of Pre-trained Language Models (PLMs) improve, there is growing con...
Despite their impressive performance in a wide range of NLP tasks, Large Language Models (LLMs) have...
Pre-trained language models reflect the inherent social biases of their training corpus. Many method...
Cheap-to-Build Very Large-Language Models (CtB-LLMs) with affordable training are emerging as the ne...
Pre-training a language model and then fine-tuning it for downstream tasks has demonstrated state-of...
Large pre-trained language models are successfully being used in a variety of tasks, across many lan...
Language model fine-tuning is essential for modern natural language processing, but is computational...
Pretrained language models can be effectively stimulated by textual prompts or demonstrations, espec...
Group bias in natural language processing tasks manifests as disparities in system error rates acros...
We introduce BitFit, a sparse-finetuning method where only the bias-terms of the model (or a subset ...
Language models (LMs) exhibit and amplify many types of undesirable biases learned from the training...
Several prior studies have suggested that word frequency biases can cause the Bert model to learn in...
Recent works have shown promising results of prompt tuning in stimulating pre-trained language model...
Warning: this paper contains model outputs exhibiting offensiveness and biases. Recently pre-trained...
When trained on large, unfiltered crawls from the Internet, language models pick up and reproduce al...