Adversarial training is the standard to train models robust against adversarial examples. However, especially for complex datasets, adversarial training incurs a significant loss in accuracy and is known to generalize poorly to stronger attacks, e.g., larger perturbations or other threat models. In this paper, we introduce confidence-calibrated adversarial training (CCAT) where the key idea is to enforce that the confidence on adversarial examples decays with their distance to the attacked examples. We show that CCAT preserves better the accuracy of normal training while robustness against adversarial examples is achieved via confidence thresholding, i.e., detecting adversarial examples based on their confidence. Most importantly, in strong...
Deep neural networks have achieved remarkable performance in various applications but are extremely ...
We present a new algorithm to train a robust malware detector. Malware is a prolific problem and mal...
Adversarial robustness has become a central goal in deep learning, both in the theory and the practi...
Despite the success of convolutional neural networks (CNNs) in many academic benchmarks for computer...
Adversarial training is an approach of increasing the robustness of models to adversarial attacks by...
Neural networks are vulnerable to adversarial attacks - small visually imperceptible crafted noise w...
Deep learning plays an important role in various disciplines, such as auto-driving, information tech...
Detecting adversarial examples currently stands as one of the biggest challenges in the field of dee...
Adversarial training, originally designed to resist test-time adversarial examples, has shown to be ...
The vulnerabilities of deep neural networks against adversarial examples have become a significant c...
Adversarial training (AT) and its variants have spearheaded progress in improving neural network rob...
Adversarial attacks are a major concern in security-centered applications, where malicious actors co...
Model Zoo (PyTorch) of non-adversarially trained models for Robust Models are less Over-Confident (N...
Image classification systems are known to be vulnerable to adversarial attacks, which are impercepti...
The adversarial training procedure proposed by Madry et al. (2018) is one of the most effective meth...
Deep neural networks have achieved remarkable performance in various applications but are extremely ...
We present a new algorithm to train a robust malware detector. Malware is a prolific problem and mal...
Adversarial robustness has become a central goal in deep learning, both in the theory and the practi...
Despite the success of convolutional neural networks (CNNs) in many academic benchmarks for computer...
Adversarial training is an approach of increasing the robustness of models to adversarial attacks by...
Neural networks are vulnerable to adversarial attacks - small visually imperceptible crafted noise w...
Deep learning plays an important role in various disciplines, such as auto-driving, information tech...
Detecting adversarial examples currently stands as one of the biggest challenges in the field of dee...
Adversarial training, originally designed to resist test-time adversarial examples, has shown to be ...
The vulnerabilities of deep neural networks against adversarial examples have become a significant c...
Adversarial training (AT) and its variants have spearheaded progress in improving neural network rob...
Adversarial attacks are a major concern in security-centered applications, where malicious actors co...
Model Zoo (PyTorch) of non-adversarially trained models for Robust Models are less Over-Confident (N...
Image classification systems are known to be vulnerable to adversarial attacks, which are impercepti...
The adversarial training procedure proposed by Madry et al. (2018) is one of the most effective meth...
Deep neural networks have achieved remarkable performance in various applications but are extremely ...
We present a new algorithm to train a robust malware detector. Malware is a prolific problem and mal...
Adversarial robustness has become a central goal in deep learning, both in the theory and the practi...