It has been shown that NLI models are usually biased with respect to the word-overlap between premise and hypothesis; they take this feature as a primary cue for predicting the entailment label. In this paper, we focus on an overlooked aspect of the overlap bias in NLI models: the reverse word-overlap bias. Our experimental results demonstrate that current NLI models are highly biased towards the non-entailment label on instances with low overlap, and the existing debiasing methods, which are reportedly successful on existing challenge datasets, are generally ineffective in addressing this category of bias. We investigate the reasons for the emergence of the overlap bias and the role of minority examples in its mitigation. For the former, w...
Natural language understanding (NLU) models often rely on dataset biases rather than intended task-r...
Group bias in natural language processing tasks manifests as disparities in system error rates acros...
Pre-trained language models reflect the inherent social biases of their training corpus. Many method...
We propose a hypothesis only baseline for diagnosing Natural Language Inference (NLI). Especially wh...
NLU models often exploit biases to achieve high dataset-specific performance without properly learni...
Natural language understanding (NLU) models tend to rely on spurious correlations (i.e., dataset bia...
We investigate how disagreement in natural language inference (NLI) annotation arises. We developed ...
Cheap-to-Build Very Large-Language Models (CtB-LLMs) with affordable training are emerging as the ne...
Datasets with significant proportions of bias present threats for training a trustworthy model on NL...
Natural Language Processing (NLP) has become increasingly utilized to provide adaptivity in educatio...
Recent studies have shown that strong Natural Language Understanding (NLU) models are prone to relyi...
Natural Language Processing (NLP) models have been found discriminative against groups of different ...
Neural network models have been very successful in natural language inference, with the best models ...
Despite their impressive performance in a wide range of NLP tasks, Large Language Models (LLMs) have...
Word embeddings have recently been shown to reflect many of the pronounced societal biases (e.g., ge...
Natural language understanding (NLU) models often rely on dataset biases rather than intended task-r...
Group bias in natural language processing tasks manifests as disparities in system error rates acros...
Pre-trained language models reflect the inherent social biases of their training corpus. Many method...
We propose a hypothesis only baseline for diagnosing Natural Language Inference (NLI). Especially wh...
NLU models often exploit biases to achieve high dataset-specific performance without properly learni...
Natural language understanding (NLU) models tend to rely on spurious correlations (i.e., dataset bia...
We investigate how disagreement in natural language inference (NLI) annotation arises. We developed ...
Cheap-to-Build Very Large-Language Models (CtB-LLMs) with affordable training are emerging as the ne...
Datasets with significant proportions of bias present threats for training a trustworthy model on NL...
Natural Language Processing (NLP) has become increasingly utilized to provide adaptivity in educatio...
Recent studies have shown that strong Natural Language Understanding (NLU) models are prone to relyi...
Natural Language Processing (NLP) models have been found discriminative against groups of different ...
Neural network models have been very successful in natural language inference, with the best models ...
Despite their impressive performance in a wide range of NLP tasks, Large Language Models (LLMs) have...
Word embeddings have recently been shown to reflect many of the pronounced societal biases (e.g., ge...
Natural language understanding (NLU) models often rely on dataset biases rather than intended task-r...
Group bias in natural language processing tasks manifests as disparities in system error rates acros...
Pre-trained language models reflect the inherent social biases of their training corpus. Many method...