A fundamental challenge of over-parameterized deep learning models is learning meaningful data representations that yield good performance on a downstream task without over-fitting spurious input features. This work proposes MaskTune, a masking strategy that prevents over-reliance on spurious (or a limited number of) features. MaskTune forces the trained model to explore new features during a single epoch finetuning by masking previously discovered features. MaskTune, unlike earlier approaches for mitigating shortcut learning, does not require any supervision, such as annotating spurious features or labels for subgroup samples in a dataset. Our empirical results on biased MNIST, CelebA, Waterbirds, and ImagenNet-9L datasets show that MaskTu...
We introduce BitFit, a sparse-finetuning method where only the bias-terms of the model (or a subset ...
Datasets with significant proportions of bias present threats for training a trustworthy model on NL...
Thesis (Ph.D.)--University of Washington, 2020Modern machine learning algorithms have been able to a...
Pre-training a language model and then fine-tuning it for downstream tasks has demonstrated state-of...
Spurious correlations in training data often lead to robustness issues since models learn to use the...
The reusability of state-of-the-art Pre-trained Language Models (PLMs) is often limited by their gen...
Many recent works indicate that the deep neural networks tend to take dataset biases as shortcuts to...
Masked language models conventionally use a masking rate of 15% due to the belief that more masking ...
While image data starts to enjoy the simple-but-effective self-supervised learning scheme built upon...
International audienceModels for fine-grained image classification tasks, where the difference betwe...
Neural network classifiers can largely rely on simple spurious features, such as backgrounds, to mak...
Neural image classifiers can often learn to make predictions by overly relying on non-predictive fea...
© 2018 Curran Associates Inc.All rights reserved. It is important to learn various types of classifi...
The selection of maskers and playback gain levels in a soundscape augmentation system is crucial to ...
Robustness has become an important consideration in deep learning. With the help of explainable AI, ...
We introduce BitFit, a sparse-finetuning method where only the bias-terms of the model (or a subset ...
Datasets with significant proportions of bias present threats for training a trustworthy model on NL...
Thesis (Ph.D.)--University of Washington, 2020Modern machine learning algorithms have been able to a...
Pre-training a language model and then fine-tuning it for downstream tasks has demonstrated state-of...
Spurious correlations in training data often lead to robustness issues since models learn to use the...
The reusability of state-of-the-art Pre-trained Language Models (PLMs) is often limited by their gen...
Many recent works indicate that the deep neural networks tend to take dataset biases as shortcuts to...
Masked language models conventionally use a masking rate of 15% due to the belief that more masking ...
While image data starts to enjoy the simple-but-effective self-supervised learning scheme built upon...
International audienceModels for fine-grained image classification tasks, where the difference betwe...
Neural network classifiers can largely rely on simple spurious features, such as backgrounds, to mak...
Neural image classifiers can often learn to make predictions by overly relying on non-predictive fea...
© 2018 Curran Associates Inc.All rights reserved. It is important to learn various types of classifi...
The selection of maskers and playback gain levels in a soundscape augmentation system is crucial to ...
Robustness has become an important consideration in deep learning. With the help of explainable AI, ...
We introduce BitFit, a sparse-finetuning method where only the bias-terms of the model (or a subset ...
Datasets with significant proportions of bias present threats for training a trustworthy model on NL...
Thesis (Ph.D.)--University of Washington, 2020Modern machine learning algorithms have been able to a...