As deep learning becomes present in many applications, we must consider possible shortcomings of these models, such as bias towards protected attributes in datasets. In this work, we focus on debiasing convolutional neural networks (CNNs), through our proposed Meta Orthogonalization algorithm. We leverage past work in debiasing word embeddings and interpretability literature to force image concepts learned by a CNN to be orthogonal to a bias direction. We empirically show through a suite of controlled bias experiments that this improves the fairness of CNNs, comparable to adversarial debiasing. We hope that this leads to new directions in debiasing and understanding deep learning modelsComputer Science
Deep learning models often learn to make predictions that rely on sensitive social attributes like g...
In image classification, debiasing aims to train a classifier to be less susceptible to dataset bias...
Important decisions are increasingly based directly on predictions from classifiers; for example, ma...
As deep learning becomes present in many applications, we must consider possible shortcomings of the...
1This work seeks to answer the question: as the (near-) orthogonality of weights is found to be a fa...
Recent discoveries have revealed that deep neural networks might behave in a biased manner in many r...
Deep Learning has achieved tremendous success in recent years in several areas such as image classif...
First we present a proof that convolutional neural networks (CNNs) with max-norm regularization, max...
Bias in classifiers is a severe issue of modern deep learning methods, especially for their applicat...
Deep neural networks (DNNs), despite their impressive ability to generalize over-capacity networks, ...
Tommasi T., Patricia N., Caputo B., Tuytelaars T., ''A deeper look at dataset bias'', 37th German co...
The presence of a bias in each image data collection has recently attracted a lot of attention in th...
Recent studies indicate that deep neural networks (DNNs) are prone to show discrimination towards ce...
Neural networks often learn to make predictions that overly rely on spurious correlation existing in...
Convolutional Neural Networks (CNNs) trained through backpropagation are central to several, competi...
Deep learning models often learn to make predictions that rely on sensitive social attributes like g...
In image classification, debiasing aims to train a classifier to be less susceptible to dataset bias...
Important decisions are increasingly based directly on predictions from classifiers; for example, ma...
As deep learning becomes present in many applications, we must consider possible shortcomings of the...
1This work seeks to answer the question: as the (near-) orthogonality of weights is found to be a fa...
Recent discoveries have revealed that deep neural networks might behave in a biased manner in many r...
Deep Learning has achieved tremendous success in recent years in several areas such as image classif...
First we present a proof that convolutional neural networks (CNNs) with max-norm regularization, max...
Bias in classifiers is a severe issue of modern deep learning methods, especially for their applicat...
Deep neural networks (DNNs), despite their impressive ability to generalize over-capacity networks, ...
Tommasi T., Patricia N., Caputo B., Tuytelaars T., ''A deeper look at dataset bias'', 37th German co...
The presence of a bias in each image data collection has recently attracted a lot of attention in th...
Recent studies indicate that deep neural networks (DNNs) are prone to show discrimination towards ce...
Neural networks often learn to make predictions that overly rely on spurious correlation existing in...
Convolutional Neural Networks (CNNs) trained through backpropagation are central to several, competi...
Deep learning models often learn to make predictions that rely on sensitive social attributes like g...
In image classification, debiasing aims to train a classifier to be less susceptible to dataset bias...
Important decisions are increasingly based directly on predictions from classifiers; for example, ma...