Universal Adversarial Attacks on Text Classifiers

Despite the vast success neural networks have achieved in different application domains, they have been proven to be vulnerable to adversarial perturbations (small changes in the input), which lead them to produce the wrong output. In this paper, we propose a novel method, based on gradient projection, for generating universal adversarial perturbations for text; namely sequence of words that can be added to any input in order to fool the classifier with high probability. We observed that text classifiers are quite vulnerable to such perturbations: inserting even a single adversarial word to the beginning of every input sequence can drop the accuracy from 93% to 50%.

Presented at:
International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Brighton, UK, 2019

 Record created 2019-02-27, last modified 2019-02-28

Rate this document:

Rate this document:
(Not yet reviewed)