Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. Universal Adversarial Attacks on Text Classifiers
 
Loading...
Thumbnail Image
conference paper

Universal Adversarial Attacks on Text Classifiers

Behjati, Melika
•
Moosavi Dezfooli, Seyed Mohsen  
•
Soleymani Baghshah, Mahdieh
Show more
2019
Proceedings of IEEE ICASSP2019 IEEE International Conference on Acoustics, Speech and Signal Processing (Icassp)
44th IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)

Despite the vast success neural networks have achieved in different application domains, they have been proven to be vulnerable to adversarial perturbations (small changes in the input), which lead them to produce the wrong output. In this paper, we propose a novel method, based on gradient projection, for generating universal adversarial perturbations for text; namely sequence of words that can be added to any input in order to fool the classifier with high probability. We observed that text classifiers are quite vulnerable to such perturbations: inserting even a single adversarial word to the beginning of every input sequence can drop the accuracy from 93% to 50%.

  • Files
  • Details
  • Metrics
Loading...
Thumbnail Image
Name

pdf.pdf

Access type

openaccess

Size

690.37 KB

Format

Adobe PDF

Checksum (MD5)

6b21d412d16668dc87fe225d625245d6

Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés