Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Journal articles
  4. Infinite-width limit of deep linear neural networks
 
research article

Infinite-width limit of deep linear neural networks

Chizat, Lenaic  
•
Colombo, Maria  
•
Fernandez-Real, Xavier  
Show more
May 6, 2024
Communications On Pure And Applied Mathematics

This paper studies the infinite-width limit of deep linear neural networks (NNs) initialized with random parameters. We obtain that, when the number of parameters diverges, the training dynamics converge (in a precise sense) to the dynamics obtained from a gradient descent on an infinitely wide deterministic linear NN. Moreover, even if the weights remain random, we get their precise law along the training dynamics, and prove a quantitative convergence result of the linear predictor in terms of the number of parameters. We finally study the continuous-time limit obtained for infinitely wide linear NNs and show that the linear predictors of the NN converge at an exponential rate to the minimal & ell;2$\ell _2$-norm minimizer of the risk.

  • Files
  • Details
  • Metrics
Loading...
Thumbnail Image
Name

Comm Pure Appl Math - 2024 - Chizat - Infinite‐width limit of deep linear neural networks.pdf

Type

Publisher

Version

Published version

Access type

openaccess

License Condition

CC BY-NC-ND

Size

1.34 MB

Format

Adobe PDF

Checksum (MD5)

68fb552cb37f136608480061db0d50cc

Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés