Named entity recognition models (NER), are widely used for identifying named entities (e.g., individuals, locations, and other information) in text documents. Machine learning based NER models are increasingly being applied in privacy-sensitive applications that need automatic and scalable identification of sensitive information to redact text for data sharing. In this paper, we study the setting when NER models are available as a black-box service for identifying sensitive information in user documents and show that these models are vulnerable to membership inference on their training datasets. With updated pre-trained NER models from spaCy, we demonstrate two distinct membership attacks on these models. Our first attack capitalizes on uni...
The Right to be Forgotten is part of the recently enacted General Data Protection Regulation (GDPR) ...
This article deals with adversarial attacks towards deep learning systems for Natural Language Proce...
We study the privacy implications of training recurrent neural networks (RNNs) with sensitive traini...
Large language models are shown to present privacy risks through memorization of training data, and ...
As in-the-wild data are increasingly involved in the training stage, machine learning applications b...
International audienceWith the rise of machine learning and data-driven models especially in the fie...
The wide adoption and application of Masked language models~(MLMs) on sensitive data (from legal to ...
Large pre-trained language models dominate the current state-of-the-art for many natural language pr...
Membership inference attacks (MIAs) against machine learning models can lead to serious privacy risk...
From fraud detection to speech recognition, including price prediction,Machine Learning (ML) applica...
Named Entity Recognition is a fundamental task in information extraction and is an essential element...
Neural network pruning has been an essential technique to reduce the computation and memory requirem...
Fine-tuning is a common and effective method for tailoring large language models (LLMs) to specializ...
Model explanations provide transparency into a trained machine learning model's blackbox behavior to...
As a long-term threat to the privacy of training data, membership inference attacks (MIAs) emerge ub...
The Right to be Forgotten is part of the recently enacted General Data Protection Regulation (GDPR) ...
This article deals with adversarial attacks towards deep learning systems for Natural Language Proce...
We study the privacy implications of training recurrent neural networks (RNNs) with sensitive traini...
Large language models are shown to present privacy risks through memorization of training data, and ...
As in-the-wild data are increasingly involved in the training stage, machine learning applications b...
International audienceWith the rise of machine learning and data-driven models especially in the fie...
The wide adoption and application of Masked language models~(MLMs) on sensitive data (from legal to ...
Large pre-trained language models dominate the current state-of-the-art for many natural language pr...
Membership inference attacks (MIAs) against machine learning models can lead to serious privacy risk...
From fraud detection to speech recognition, including price prediction,Machine Learning (ML) applica...
Named Entity Recognition is a fundamental task in information extraction and is an essential element...
Neural network pruning has been an essential technique to reduce the computation and memory requirem...
Fine-tuning is a common and effective method for tailoring large language models (LLMs) to specializ...
Model explanations provide transparency into a trained machine learning model's blackbox behavior to...
As a long-term threat to the privacy of training data, membership inference attacks (MIAs) emerge ub...
The Right to be Forgotten is part of the recently enacted General Data Protection Regulation (GDPR) ...
This article deals with adversarial attacks towards deep learning systems for Natural Language Proce...
We study the privacy implications of training recurrent neural networks (RNNs) with sensitive traini...