Existing NLP datasets contain various biases that models can easily exploit to achieve high performances on the corresponding evaluation sets. However, focusing on dataset-specific biases limits their ability to learn more generalizable knowledge about the task from more general data patterns. In this paper, we investigate the impact of debiasing methods for improving generalization and propose a general framework for improving the performance on both in-domain and out-of-domain datasets by concurrent modeling of multiple biases in the training data. Our framework weights each example based on the biases it contains and the strength of those biases in the training data. It then uses these weights in the training objective so that the model ...
Past work that investigates out-of-domain performance of QA systems has mainly focused on general do...
Cheap-to-Build Very Large-Language Models (CtB-LLMs) with affordable training are emerging as the ne...
Language Representation Models (LRMs) trained with real-world data may capture and exacerbate undesi...
NLU models often exploit biases to achieve high dataset-specific performance without properly learni...
Models of various NLP tasks have been shown to exhibit stereotypes, and the bias in the question ans...
Deep Learning has achieved tremendous success in recent years in several areas such as image classif...
Neural networks often make predictions relying on the spurious correlations from the datasets rather...
Thesis (Ph.D.)--University of Washington, 2020Modern machine learning algorithms have been able to a...
Natural language understanding (NLU) models often rely on dataset biases rather than intended task-r...
Generalization beyond the training distribution is a core challenge in machine learning. The common ...
Recent discoveries have revealed that deep neural networks might behave in a biased manner in many r...
As the representation capability of Pre-trained Language Models (PLMs) improve, there is growing con...
When trained on large, unfiltered crawls from the Internet, language models pick up and reproduce al...
Deep learning models often learn to make predictions that rely on sensitive social attributes like g...
Recent studies have shown that strong Natural Language Understanding (NLU) models are prone to relyi...
Past work that investigates out-of-domain performance of QA systems has mainly focused on general do...
Cheap-to-Build Very Large-Language Models (CtB-LLMs) with affordable training are emerging as the ne...
Language Representation Models (LRMs) trained with real-world data may capture and exacerbate undesi...
NLU models often exploit biases to achieve high dataset-specific performance without properly learni...
Models of various NLP tasks have been shown to exhibit stereotypes, and the bias in the question ans...
Deep Learning has achieved tremendous success in recent years in several areas such as image classif...
Neural networks often make predictions relying on the spurious correlations from the datasets rather...
Thesis (Ph.D.)--University of Washington, 2020Modern machine learning algorithms have been able to a...
Natural language understanding (NLU) models often rely on dataset biases rather than intended task-r...
Generalization beyond the training distribution is a core challenge in machine learning. The common ...
Recent discoveries have revealed that deep neural networks might behave in a biased manner in many r...
As the representation capability of Pre-trained Language Models (PLMs) improve, there is growing con...
When trained on large, unfiltered crawls from the Internet, language models pick up and reproduce al...
Deep learning models often learn to make predictions that rely on sensitive social attributes like g...
Recent studies have shown that strong Natural Language Understanding (NLU) models are prone to relyi...
Past work that investigates out-of-domain performance of QA systems has mainly focused on general do...
Cheap-to-Build Very Large-Language Models (CtB-LLMs) with affordable training are emerging as the ne...
Language Representation Models (LRMs) trained with real-world data may capture and exacerbate undesi...