Adversarial attacks cause machine learning models to produce wrong predictions by minimally perturbing their input. In this thesis, we take a step towards understanding how these perturbations affect the intermediate data representations of the model. Specifically, we compare standard and adversarial representations for models of varying robustness based on a variety of similarity metrics. In fact, we find that it’s possible to detect adversarial examples by examining nearby examples, though we also find that this method can be circumvented by an adaptive attack. We then explore methods to improve generalization to natural distribution shift and hypothesize that models trained with different notions of feature bias will learn fundamentally ...
In recent years, machine learning (ML) models have been extensively used in software analytics, such...
International audienceRecent research has shown that machine learning systems, including state-of-th...
Despite impressive success in many tasks, deep learning models are shown to rely on spurious feature...
Machine Learning (ML) models are vulnerable to adversarial samples — human imperceptible changes to ...
The reason for the existence of adversarial samples is still barely understood. Here, we explore the...
© 2019 Neural information processing systems foundation. All rights reserved. Adversarial examples h...
In this thesis we explore adversarial examples for simple model families and simple data distributio...
Deep learning plays an important role in various disciplines, such as auto-driving, information tech...
International audienceDeep learning classifiers are now known to have flaws in the representations o...
Modern machine learning models can be difficult to probe and understand after they have been trained...
Neural language models show vulnerability to adversarial examples which are semantically similar to ...
Machine learning is used in myriad aspects, both in academic research and in everyday life, includin...
Over the last decade, machine learning systems have achieved state-of-the-art performance in many fi...
Prompted by its performance on a variety of benchmark tasks, machine learning (ML) is now being appl...
Deep learning has been a popular topic and has achieved success in many areas. It has drawn the atte...
In recent years, machine learning (ML) models have been extensively used in software analytics, such...
International audienceRecent research has shown that machine learning systems, including state-of-th...
Despite impressive success in many tasks, deep learning models are shown to rely on spurious feature...
Machine Learning (ML) models are vulnerable to adversarial samples — human imperceptible changes to ...
The reason for the existence of adversarial samples is still barely understood. Here, we explore the...
© 2019 Neural information processing systems foundation. All rights reserved. Adversarial examples h...
In this thesis we explore adversarial examples for simple model families and simple data distributio...
Deep learning plays an important role in various disciplines, such as auto-driving, information tech...
International audienceDeep learning classifiers are now known to have flaws in the representations o...
Modern machine learning models can be difficult to probe and understand after they have been trained...
Neural language models show vulnerability to adversarial examples which are semantically similar to ...
Machine learning is used in myriad aspects, both in academic research and in everyday life, includin...
Over the last decade, machine learning systems have achieved state-of-the-art performance in many fi...
Prompted by its performance on a variety of benchmark tasks, machine learning (ML) is now being appl...
Deep learning has been a popular topic and has achieved success in many areas. It has drawn the atte...
In recent years, machine learning (ML) models have been extensively used in software analytics, such...
International audienceRecent research has shown that machine learning systems, including state-of-th...
Despite impressive success in many tasks, deep learning models are shown to rely on spurious feature...