International audienceModel explanations provide transparency into a trained machine learning model’s blackbox behavior to a model builder. They indicate the influence of different input attributes to its corresponding model prediction. The dependency of explanations on input raises privacy concerns for sensitive user data. However, current literature has limited discussion on privacy risks of model explanations. We focus on the specific privacy risk of attribute inference attack wherein an adversary infers sensitive attributes of an input (e.g., Race and Sex) given its model explanations. We design the first attribute inference attack against model explanations in two threat models where model builder either (a) includes the sensitive attr...
It is known that deep neural networks, trained for the classification of non-sensitive target attrib...
Both researchers and industry have increased their employ of machine learning in new applications wi...
It is observed in the literature that data augmentation can significantly mitigate membership infere...
Model explanations provide transparency into a trained machine learning model's blackbox behavior to...
Machine learning (ML) models have been deployed for high-stakes applications. Due to class imbalance...
Transparency of algorithmic systems is an important area of research, which has been discussed as a ...
The wide adoption and application of Masked language models~(MLMs) on sensitive data (from legal to ...
Inference attacks against Machine Learning (ML) models allow adversaries to learn information about ...
Large-scale pre-trained models are increasingly adapted to downstream tasks through a new paradigm c...
Fine-tuning is a common and effective method for tailoring large language models (LLMs) to specializ...
Machine learning (ML) has become a core component of many real-world applications and training data ...
Recent development in the field of explainable artificial intelligence (XAI) has helped improve trus...
Machine learning (ML) has been widely adopted in various privacy-critical applications, e.g., face r...
We investigate an attack on a machine learning model that predicts whether a person or household wil...
Machine learning models are often trained on sensitive and proprietary datasets. Yet what -- and und...
It is known that deep neural networks, trained for the classification of non-sensitive target attrib...
Both researchers and industry have increased their employ of machine learning in new applications wi...
It is observed in the literature that data augmentation can significantly mitigate membership infere...
Model explanations provide transparency into a trained machine learning model's blackbox behavior to...
Machine learning (ML) models have been deployed for high-stakes applications. Due to class imbalance...
Transparency of algorithmic systems is an important area of research, which has been discussed as a ...
The wide adoption and application of Masked language models~(MLMs) on sensitive data (from legal to ...
Inference attacks against Machine Learning (ML) models allow adversaries to learn information about ...
Large-scale pre-trained models are increasingly adapted to downstream tasks through a new paradigm c...
Fine-tuning is a common and effective method for tailoring large language models (LLMs) to specializ...
Machine learning (ML) has become a core component of many real-world applications and training data ...
Recent development in the field of explainable artificial intelligence (XAI) has helped improve trus...
Machine learning (ML) has been widely adopted in various privacy-critical applications, e.g., face r...
We investigate an attack on a machine learning model that predicts whether a person or household wil...
Machine learning models are often trained on sensitive and proprietary datasets. Yet what -- and und...
It is known that deep neural networks, trained for the classification of non-sensitive target attrib...
Both researchers and industry have increased their employ of machine learning in new applications wi...
It is observed in the literature that data augmentation can significantly mitigate membership infere...