© Springer Nature Switzerland AG 2020. Recently, generating adversarial examples has become an important means of measuring robustness of a deep learning model. Adversarial examples help us identify the susceptibilities of the model and further counter those vulnerabilities by applying adversarial training techniques. In natural language domain, small perturbations in the form of misspellings or paraphrases can drastically change the semantics of the text. We propose a reinforcement learning based approach towards generating adversarial examples in black-box settings. We demonstrate that our method is able to fool well-trained models for (a) IMDB sentiment classification task and (b) AG’s news corpus news categorization task with significan...
Adversarial examples are inputs to a machine learning system that result in an incorrect output from...
In adversarial attacks intended to confound deep learning models, most studies have focused on limit...
Adversarial training is an approach of increasing the robustness of models to adversarial attacks by...
In recent years, the neural networks are widely used in image processing, natural language processin...
We study an important and challenging task of attacking natural language processing models in a hard...
Black-box attacks in deep reinforcement learning usually retrain substitute policies to mimic behavi...
The monumental achievements of deep learning (DL) systems seem to guarantee the absolute superiority...
Neural language models show vulnerability to adversarial examples which are semantically similar to ...
Research shows that natural language processing models are generally considered to be vulnerable to ...
Text classification is a basic task in natural language processing, but the small character perturba...
Generating adversarial examples for natural language is hard, as natural language consists of discre...
In the thesis, we explore the prospects of creating adversarial examples using various generative mo...
Adversarial examples are helpful for analyzing and improving the robustness of text classifiers. Gen...
We study an important task of attacking natural language processing models in a black box setting. W...
We propose a simple and general method to regularize the fine-tuning of Transformer-based encoders f...
Adversarial examples are inputs to a machine learning system that result in an incorrect output from...
In adversarial attacks intended to confound deep learning models, most studies have focused on limit...
Adversarial training is an approach of increasing the robustness of models to adversarial attacks by...
In recent years, the neural networks are widely used in image processing, natural language processin...
We study an important and challenging task of attacking natural language processing models in a hard...
Black-box attacks in deep reinforcement learning usually retrain substitute policies to mimic behavi...
The monumental achievements of deep learning (DL) systems seem to guarantee the absolute superiority...
Neural language models show vulnerability to adversarial examples which are semantically similar to ...
Research shows that natural language processing models are generally considered to be vulnerable to ...
Text classification is a basic task in natural language processing, but the small character perturba...
Generating adversarial examples for natural language is hard, as natural language consists of discre...
In the thesis, we explore the prospects of creating adversarial examples using various generative mo...
Adversarial examples are helpful for analyzing and improving the robustness of text classifiers. Gen...
We study an important task of attacking natural language processing models in a black box setting. W...
We propose a simple and general method to regularize the fine-tuning of Transformer-based encoders f...
Adversarial examples are inputs to a machine learning system that result in an incorrect output from...
In adversarial attacks intended to confound deep learning models, most studies have focused on limit...
Adversarial training is an approach of increasing the robustness of models to adversarial attacks by...