Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. A Catalyst Framework for Minimax Optimization
 
conference paper

A Catalyst Framework for Minimax Optimization

Yang, Junchi
•
Zhang, Siqi
•
Kiyavash, Negar  
Show more
January 1, 2020
Advances In Neural Information Processing Systems (Neurips 2020)
34th Conference on Neural Information Processing Systems (NeurIPS)

We introduce a generic two-loop scheme for smooth minimax optimization with strongly-convex-concave objectives. Our approach applies the accelerated proximal point framework (or Catalyst) to the associated dual problem and takes full advantage of existing gradient-based algorithms to solve a sequence of well-balanced strongly-convex-strongly-concave minimax problems. Despite its simplicity, this leads to a family of near-optimal algorithms with improved complexity over all existing methods designed for strongly-convex-concave minimax problems. Additionally, we obtain the first variance-reduced algorithms for this class of minimax problems with finite-sum structure and establish faster convergence rate than batch algorithms. Furthermore, when extended to the nonconvex-concave minimax optimization, our algorithm again achieves the state-of-the-art complexity for finding a stationary point. We carry out several numerical experiments showcasing the superiority of the Catalyst framework in practice.

  • Files
  • Details
  • Metrics
Type
conference paper
Web of Science ID

WOS:000627697000011

Author(s)
Yang, Junchi
Zhang, Siqi
Kiyavash, Negar  
He, Niao
Date Issued

2020-01-01

Publisher

NEURAL INFORMATION PROCESSING SYSTEMS (NIPS)

Publisher place

La Jolla

Published in
Advances In Neural Information Processing Systems (Neurips 2020)
Series title/Series vol.

Advances in Neural Information Processing Systems

Volume

33

Subjects

Computer Science, Artificial Intelligence

•

Computer Science, Information Systems

•

Computer Science

•

variational-inequalities

•

monotone-operators

•

convergence

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
BAN  
Event nameEvent placeEvent date
34th Conference on Neural Information Processing Systems (NeurIPS)

ELECTR NETWORK

Dec 06-12, 2020

Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/177586
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés