NLU models often exploit biases to achieve high dataset-specific performance without properly learning the intended task. Recently proposed debiasing methods are shown to be effective in mitigating this tendency. However, these methods rely on a major assumption that the types of bias should be known a-priori, which limits their application to many NLU tasks and datasets. In this work, we present the first step to bridge this gap by introducing a self-debiasing framework that prevents models from mainly utilizing biases without knowing them in advance. The proposed framework is general and complementary to the existing debiasing methods. We show that it allows these existing methods to retain the improvement on the challenge datasets (i.e.,...
It has been shown that NLI models are usually biased with respect to the word-overlap between premis...
Cheap-to-Build Very Large-Language Models (CtB-LLMs) with affordable training are emerging as the ne...
Models notoriously suffer from dataset biases which are detrimental to robustness and generalization...
Natural language understanding (NLU) models often rely on dataset biases rather than intended task-r...
Recent studies have shown that strong Natural Language Understanding (NLU) models are prone to relyi...
When trained on large, unfiltered crawls from the Internet, language models pick up and reproduce al...
Existing NLP datasets contain various biases that models can easily exploit to achieve high performa...
Datasets with significant proportions of bias present threats for training a trustworthy model on NL...
Recent discoveries have revealed that deep neural networks might behave in a biased manner in many r...
Improperly constructed datasets can result in inaccurate inferences. For instance, models trained on...
Deep Learning has achieved tremendous success in recent years in several areas such as image classif...
Language Representation Models (LRMs) trained with real-world data may capture and exacerbate undesi...
In image classification, debiasing aims to train a classifier to be less susceptible to dataset bias...
Thesis (Ph.D.)--University of Washington, 2020Modern machine learning algorithms have been able to a...
Deep neural networks (DNNs), despite their impressive ability to generalize over-capacity networks, ...
It has been shown that NLI models are usually biased with respect to the word-overlap between premis...
Cheap-to-Build Very Large-Language Models (CtB-LLMs) with affordable training are emerging as the ne...
Models notoriously suffer from dataset biases which are detrimental to robustness and generalization...
Natural language understanding (NLU) models often rely on dataset biases rather than intended task-r...
Recent studies have shown that strong Natural Language Understanding (NLU) models are prone to relyi...
When trained on large, unfiltered crawls from the Internet, language models pick up and reproduce al...
Existing NLP datasets contain various biases that models can easily exploit to achieve high performa...
Datasets with significant proportions of bias present threats for training a trustworthy model on NL...
Recent discoveries have revealed that deep neural networks might behave in a biased manner in many r...
Improperly constructed datasets can result in inaccurate inferences. For instance, models trained on...
Deep Learning has achieved tremendous success in recent years in several areas such as image classif...
Language Representation Models (LRMs) trained with real-world data may capture and exacerbate undesi...
In image classification, debiasing aims to train a classifier to be less susceptible to dataset bias...
Thesis (Ph.D.)--University of Washington, 2020Modern machine learning algorithms have been able to a...
Deep neural networks (DNNs), despite their impressive ability to generalize over-capacity networks, ...
It has been shown that NLI models are usually biased with respect to the word-overlap between premis...
Cheap-to-Build Very Large-Language Models (CtB-LLMs) with affordable training are emerging as the ne...
Models notoriously suffer from dataset biases which are detrimental to robustness and generalization...