Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. What can linearized neural networks actually say about generalization?
 
conference paper

What can linearized neural networks actually say about generalization?

Ortiz Jimenez, Guillermo  
•
Moosavi Dezfooli, Seyed Mohsen  
•
Frossard, Pascal  
Ranzato, M.
•
Beygelzimer, A.
Show more
December 6, 2021
Advances in Neural Information Processing Systems 34
Thirty-fifth Conference on Neural Information Processing Systems (NeurIPS 2021)

For certain infinitely-wide neural networks, the neural tangent kernel (NTK) theory fully characterizes generalization, but for the networks used in practice, the empirical NTK only provides a rough first-order approximation. Still, a growing body of work keeps leveraging this approximation to successfully analyze important deep learning phenomena and design algorithms for new applications. In our work, we provide strong empirical evidence to determine the practical validity of such approximation by conducting a systematic comparison of the behavior of different neural networks and their linear approximations on different tasks. We show that the linear approximations can indeed rank the learning complexity of certain tasks for neural networks, even when they achieve very different performances. However, in contrast to what was previously reported, we discover that neural networks do not always perform better than their kernel approximations, and reveal that the performance gap heavily depends on architecture, dataset size and training task. We discover that networks overfit to these tasks mostly due to the evolution of their kernel during training, thus, revealing a new type of implicit bias.

  • Files
  • Details
  • Metrics
Loading...
Thumbnail Image
Name

2106.06770(1).pdf

Type

Publisher's Version

Version

Published version

Access type

openaccess

License Condition

CC BY

Size

2.24 MB

Format

Adobe PDF

Checksum (MD5)

80add4cd2701809ab1fa39c1d9bbd567

Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés