Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Books and Book parts
  4. Robust and Private Federated Learning on LLMs
 
book part or chapter

Robust and Private Federated Learning on LLMs

Guerraoui, Rachid  
•
Gupta, Nirupam  
January 1, 2024
Large Language Models in Cybersecurity: Threats, Exposure and Mitigation

Large Language Models (LLMs) have gained significant attention in recent years due to their potential to revolutionize various industries and sectors. However, scaling LLMs further requires access to substantial linguistic resources that are being rapidly depleted. Moreover, the available text sources such as emails, social media interactions, or internal documents may contain private information, making them susceptible to misuse. On-premises Federated Learning (FL) with privacy-preserving model updates is an alternative avenue for LLMs’ development that ensures data sovereignty and enables peers to collaborate while ensuring that the sensitive parts of their private data cannot be reconstructed. However, in the case of large-scale FL, there is also a risk of malicious users attempting to poison LLMs for their benefit. The problem of protecting the learning procedure against such users is known as Byzantine-robustness, and it is crucial to develop models that perform accurately despite faulty machines and poisonous data. Designing FL methods that are simultaneously privacy-preserving and Byzantine-robust is challenging. However, ongoing research suggests ways to incorporate the differentially-private Gaussian mechanism for privacy preservation and spectral robust-averaging for robustness. However, whether this approach applies to LLMs or whether a major player in the domain would emerge and capture all private information sources through network effects remains to be seen.

  • Files
  • Details
  • Metrics
Type
book part or chapter
DOI
10.1007/978-3-031-54827-7_21
Scopus ID

2-s2.0-85207237225

Author(s)
Guerraoui, Rachid  

EPFL

Gupta, Nirupam  

EPFL

Date Issued

2024-01-01

Publisher

Springer Nature

Publisher place

Chams (Switzerland)

Published in
Large Language Models in Cybersecurity: Threats, Exposure and Mitigation
DOI of the book
https://doi.org/10.1007/978-3-031-54827-7
ISBN of the book

9783031548277

9783031548260

Total of pages

249

Start page

189

End page

196

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
DCL  
Available on Infoscience
January 27, 2025
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/245293
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés