Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. Implicit Gradient Alignment in Distributed and Federated Learning
 
conference paper

Implicit Gradient Alignment in Distributed and Federated Learning

Dandi, Yatin  
•
Barba, Luis  
•
Jaggi, Martin  
January 1, 2022
Thirty-Sixth Aaai Conference On Artificial Intelligence / Thirty-Fourth Conference On Innovative Applications Of Artificial Intelligence / The Twelveth Symposium On Educational Advances In Artificial Intelligence
36th AAAI Conference on Artificial Intelligence / 34th Conference on Innovative Applications of Artificial Intelligence / 12th Symposium on Educational Advances in Artificial Intelligence

A major obstacle to achieving global convergence in distributed and federated learning is the misalignment of gradients across clients, or mini-batches due to heterogeneity and stochasticity of the distributed data. In this work, we show that data heterogeneity can in fact be exploited to improve generalization performance through implicit regularization. One way to alleviate the effects of heterogeneity is to encourage the alignment of gradients across different clients throughout training. Our analysis reveals that this goal can be accomplished by utilizing the right optimization method that replicates the implicit regularization effect of SGD, leading to gradient alignment as well as improvements in test accuracies. Since the existence of this regularization in SGD completely relies on the sequential use of different mini-batches during training, it is inherently absent when training with large mini-batches. To obtain the generalization benefits of this regularization while increasing parallelism, we propose a novel GradAlign algorithm that induces the same implicit regularization while allowing the use of arbitrarily large batches in each update. We experimentally validate the benefits of our algorithm in different distributed and federated learning settings.

  • Details
  • Metrics
Type
conference paper
DOI
10.1609/aaai.v36i6.20597
Web of Science ID

WOS:000893636206063

Author(s)
Dandi, Yatin  
Barba, Luis  
Jaggi, Martin  
Date Issued

2022-01-01

Publisher

ASSOC ADVANCEMENT ARTIFICIAL INTELLIGENCE

Publisher place

Palo Alto

Published in
Thirty-Sixth Aaai Conference On Artificial Intelligence / Thirty-Fourth Conference On Innovative Applications Of Artificial Intelligence / The Twelveth Symposium On Educational Advances In Artificial Intelligence
ISBN of the book

978-1-57735-876-3

Series title/Series vol.

AAAI Conference on Artificial Intelligence

Start page

6454

End page

6462

Subjects

Computer Science, Artificial Intelligence

•

Computer Science

•

networks

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

Event nameEvent placeEvent date
36th AAAI Conference on Artificial Intelligence / 34th Conference on Innovative Applications of Artificial Intelligence / 12th Symposium on Educational Advances in Artificial Intelligence

ELECTR NETWORK

Feb 22-Mar 01, 2022

Available on Infoscience
February 27, 2023
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/195193
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés