Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Journal articles
  4. Fast Adversarial Training With Adaptive Step Size
 
research article

Fast Adversarial Training With Adaptive Step Size

Huang, Zhichao
•
Fan, Yanbo
•
Liu, Chen
Show more
January 1, 2023
Ieee Transactions On Image Processing

While adversarial training and its variants have shown to be the most effective algorithms to defend against adversarial attacks, their extremely slow training process makes it hard to scale to large datasets like ImageNet. The key idea of recent works to accelerate adversarial training is to substitute multi-step attacks (e.g., PGD) with single-step attacks (e.g., FGSM). However, these single-step methods suffer from catastrophic overfitting, where the accuracy against PGD attack suddenly drops to nearly 0% during training, and the network totally loses its robustness. In this work, we study the phenomenon from the perspective of training instances. We show that catastrophic overfitting is instance-dependent, and fitting instances with larger input gradient norm is more likely to cause catastrophic overfitting. Based on our findings, we propose a simple but effective method, Adversarial Training with Adaptive Step size (ATAS). ATAS learns an instance-wise adaptive step size that is inversely proportional to its gradient norm. Our theoretical analysis shows that ATAS converges faster than the commonly adopted non-adaptive counterparts. Empirically, ATAS consistently mitigates catastrophic overfitting and achieves higher robust accuracy on CIFAR10, CIFAR100, and ImageNet when evaluated on various adversarial budgets. Our code is released at https://github.com/HuangZhiChao95/ATAS.

  • Details
  • Metrics
Type
research article
DOI
10.1109/TIP.2023.3326398
Web of Science ID

WOS:001123333700009

Author(s)
Huang, Zhichao
Fan, Yanbo
Liu, Chen
Zhang, Weizhong
Zhang, Yong
Salzmann, Mathieu  
Susstrunk, Sabine
Wang, Jue
Date Issued

2023-01-01

Publisher

Ieee-Inst Electrical Electronics Engineers Inc

Published in
Ieee Transactions On Image Processing
Volume

32

Start page

6102

End page

6114

Subjects

Technology

•

Training

•

Robustness

•

Perturbation Methods

•

Optimization

•

Fans

•

Stochastic Processes

•

Standards

•

Adversarial Examples

•

Fast Adversarial Training

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
CVLAB  
FunderGrant Number

CityU APRC

Available on Infoscience
February 20, 2024
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/204757
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés