A line of work has shown that natural text processing models are vulnerable to adversarial examples. Correspondingly, various defense methods are proposed to mitigate the threat of textual adversarial examples, eg, adversarial training, input transformations, detection, etc. In this work, we treat the optimization process for synonym substitution based textual adversarial attacks as a specific sequence of word replacement, in which each word mutually influences other words. We identify that we could destroy such mutual interaction and eliminate the adversarial perturbation by randomly substituting a word with its synonyms. Based on this observation, we propose a novel textual adversarial example detection method, termed Randomized Substitut...
Neural ranking models (NRMs) have shown remarkable success in recent years, especially with pre-trai...
For humans, distinguishing machine generated text from human written text is men- tally taxing and s...
Text classification is a basic task in natural language processing, but the small character perturba...
Research shows that natural language processing models are generally considered to be vulnerable to ...
Modern text classification models are susceptible to adversarial examples, perturbed versions of the...
Adversarial attacks in NLP challenge the way we look at language models. The goal of this kind of ad...
Deep learning models have excelled in solving many problems in Natural Language Processing, but are ...
In recent years, the neural networks are widely used in image processing, natural language processin...
With the advent of high-performance computing devices, deep neural networks have gained a lot of pop...
We present three large-scale experiments on binary text matching classification task both in Chinese...
Adversarial training is the most empirically successful approach in improving the robustness of deep...
NLP researchers propose different word-substitute black-box attacks that can fool text classificatio...
Neural language models show vulnerability to adversarial examples which are semantically similar to ...
Despite deep neural networks (DNNs) having achieved impressive performance in various domains, it ha...
Adversarial examples in NLP are receiving increasing research attention. One line of investigation i...
Neural ranking models (NRMs) have shown remarkable success in recent years, especially with pre-trai...
For humans, distinguishing machine generated text from human written text is men- tally taxing and s...
Text classification is a basic task in natural language processing, but the small character perturba...
Research shows that natural language processing models are generally considered to be vulnerable to ...
Modern text classification models are susceptible to adversarial examples, perturbed versions of the...
Adversarial attacks in NLP challenge the way we look at language models. The goal of this kind of ad...
Deep learning models have excelled in solving many problems in Natural Language Processing, but are ...
In recent years, the neural networks are widely used in image processing, natural language processin...
With the advent of high-performance computing devices, deep neural networks have gained a lot of pop...
We present three large-scale experiments on binary text matching classification task both in Chinese...
Adversarial training is the most empirically successful approach in improving the robustness of deep...
NLP researchers propose different word-substitute black-box attacks that can fool text classificatio...
Neural language models show vulnerability to adversarial examples which are semantically similar to ...
Despite deep neural networks (DNNs) having achieved impressive performance in various domains, it ha...
Adversarial examples in NLP are receiving increasing research attention. One line of investigation i...
Neural ranking models (NRMs) have shown remarkable success in recent years, especially with pre-trai...
For humans, distinguishing machine generated text from human written text is men- tally taxing and s...
Text classification is a basic task in natural language processing, but the small character perturba...