Natural language models and systems have been shown to reflect gender bias existing in training data. This bias can impact on the downstream task that machine learning models, built on this training data, are to accomplish. A variety of techniques have been proposed to mitigate gender bias in training data. In this paper we compare different gender bias mitigation approaches on a classification task. We consider mitigation techniques that manipulate the training data itself, including data scrubbing, gender swapping and counterfactual data augmentation approaches. We also look at using de-biased word embeddings in the representation of the training data. We evaluate the effectiveness of the different approaches at reducing the gender bias in the...
Contextualized word embeddings have been replacing standard embeddings as the representational knowl...
Contextualized word embeddings have been replacing standard embeddings as the representational knowl...
Contextualized word embeddings have been replacing standard embeddings as the representational knowl...
Predictions from machine learning models can reflect biases in the data on which they are trained. G...
International Workshop on Algorithmic Bias in Search and Recommendation (Bias 2020), Lisbon, Portuga...
Gender bias in Natural Language Processing (NLP) models is a non-trivial problem that can perpetuate...
This paper proposes two intuitive metrics, skew and stereotype, that quantify and analyse the gender...
This paper proposes two intuitive metrics, skew and stereotype, that quantify and analyse the gender...
This paper proposes two intuitive metrics, skew and stereotype, that quantify and analyse the gender...
This paper proposes two intuitive metrics, skew and stereotype, that quantify and analyse the gender...
Pre-trained language models encode undesirable social biases, which are further exacerbated in downs...
Language models (LMs) exhibit and amplify many types of undesirable biases learned from the training...
Language models are used for a variety of downstream applications, such as improving web search resu...
Language models are used for a variety of downstream applications, such as improving web search resu...
Language models are used for a variety of downstream applications, such as improving web search resu...
Contextualized word embeddings have been replacing standard embeddings as the representational knowl...
Contextualized word embeddings have been replacing standard embeddings as the representational knowl...
Contextualized word embeddings have been replacing standard embeddings as the representational knowl...
Predictions from machine learning models can reflect biases in the data on which they are trained. G...
International Workshop on Algorithmic Bias in Search and Recommendation (Bias 2020), Lisbon, Portuga...
Gender bias in Natural Language Processing (NLP) models is a non-trivial problem that can perpetuate...
This paper proposes two intuitive metrics, skew and stereotype, that quantify and analyse the gender...
This paper proposes two intuitive metrics, skew and stereotype, that quantify and analyse the gender...
This paper proposes two intuitive metrics, skew and stereotype, that quantify and analyse the gender...
This paper proposes two intuitive metrics, skew and stereotype, that quantify and analyse the gender...
Pre-trained language models encode undesirable social biases, which are further exacerbated in downs...
Language models (LMs) exhibit and amplify many types of undesirable biases learned from the training...
Language models are used for a variety of downstream applications, such as improving web search resu...
Language models are used for a variety of downstream applications, such as improving web search resu...
Language models are used for a variety of downstream applications, such as improving web search resu...
Contextualized word embeddings have been replacing standard embeddings as the representational knowl...
Contextualized word embeddings have been replacing standard embeddings as the representational knowl...
Contextualized word embeddings have been replacing standard embeddings as the representational knowl...