Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. The Hidden Vulnerability of Distributed Learning in Byzantium
 
conference paper not in proceedings

The Hidden Vulnerability of Distributed Learning in Byzantium

El Mhamdi, El Mahdi  
•
Guerraoui, Rachid  
•
Rouault, Sébastien Louis Alexandre  
2018
International Conference on Machine Learning

While machine learning is going through an era of celebrated success, concerns have been raised about the vulnerability of its backbone: stochastic gradient descent (SGD). Recent approaches have been proposed to ensure the robustness of distributed SGD against adversarial (Byzantine) workers sending \emph{poisoned} gradients during the training phase. Some of these approaches have been proven \emph{Byzantine--resilient}: they ensure the \emph{convergence} of SGD despite the presence of a minority of adversarial workers. We show in this paper that \emph{convergence is not enough}. In high dimension $d \gg 1$, an adver-sary can build on the loss function's non--convexity to make SGD converge to \emph{ineffective} models. More precisely, we bring to light that existing Byzantine--resilient schemes leave a \emph{margin of poisoning} of $\bigOmega\left(f(d)\right)$, where $f(d)$ increases at least like $\sqrt{d}$. Based on this \emph{leeway}, we build a simple attack, and experimentally show its strong to utmost effectivity on CIFAR--10 and MNIST. We introduce \emph{Bulyan}, and prove it significantly reduces the attacker's leeway to a narrow $\bigO,( \sfrac{1}{\sqrt{d~}})$ bound. We empirically show that Bulyan does not suffer the fragility of existing aggregation rules and, at a reasonable cost in terms of required batch size, achieves convergence \emph{as if} only non--Byzantine gradients had been used to update the model.

  • Files
  • Details
  • Metrics
Type
conference paper not in proceedings
Author(s)
El Mhamdi, El Mahdi  
Guerraoui, Rachid  
Rouault, Sébastien Louis Alexandre  
Date Issued

2018

Total of pages

13

Subjects

Machine Learning

•

Distributed Algorithms

•

Byzantine fault tolerance

•

Robustness

•

stochastic gradient descent

•

SGD

•

Poisoning attack

•

adversarial machine learning

•

ml-ai

Note

camera ready version available also on ICML proceedings (open access)

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
DCL  
Event nameEvent placeEvent date
International Conference on Machine Learning

Stockholm, Sweden

July 10-15, 2018

Available on Infoscience
July 17, 2018
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/147400
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés