Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. The Gaussian equivalence of generative models for learning with shallow neural networks
 
conference paper

The Gaussian equivalence of generative models for learning with shallow neural networks

Goldt, Sebastian
•
Loureiro, Bruno  
•
Reeves, Galen
Show more
December 16, 2021
Proceedings of the 2nd Mathematical and Scientific Machine Learning Conference
2nd Mathematical and Scientific Machine Learning Conference

Understanding the impact of data structure on the computational tractability of learning is a key challenge for the theory of neural networks. Many theoretical works do not explicitly model training data, or assume that inputs are drawn component-wise independently from some simple probability distribution. Here, we go beyond this simple paradigm by studying the performance of neural networks trained on data drawn from pre-trained generative models. This is possible due to a Gaussian equivalence stating that the key metrics of interest, such as the training and test errors, can be fully captured by an appropriately chosen Gaussian model. We provide three strands of rigorous, analytical and numerical evidence corroborating this equivalence. First, we establish rigorous conditions for the Gaussian equivalence to hold in the case of single-layer generative models, as well as deterministic rates for convergence in distribution. Second, we leverage this equivalence to derive a closed set of equations describing the generalisation performance of two widely studied machine learning problems: two-layer neural networks trained using one-pass stochastic gradient descent, and full-batch pre-learned features or kernel methods. Finally, we perform experiments demonstrating how our theory applies to deep, pre-trained generative models. These results open a viable path to the theoretical study of machine learning models with realistic data. Keywords: Neural networks, Generative models, Stochastic Gradient Descent, Random Features.

  • Files
  • Details
  • Metrics
Type
conference paper
Author(s)
Goldt, Sebastian
Loureiro, Bruno  
Reeves, Galen
Krzakala, Florent  
Mezard, Marc
Zdeborová, Lenka  
Date Issued

2021-12-16

Published in
Proceedings of the 2nd Mathematical and Scientific Machine Learning Conference
Total of pages

426-471

Series title/Series vol.

Proceedings of Machine Learning Research; 145

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
SPOC1  
SPOC2  
IDEPHICS1  
Show more
Event nameEvent placeEvent date
2nd Mathematical and Scientific Machine Learning Conference

Online

August 16-19, 2021

Available on Infoscience
December 22, 2022
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/193488
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés