Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. Generalization Properties of NAS under Activation and Skip Connection Search
 
conference paper

Generalization Properties of NAS under Activation and Skip Connection Search

Zhu, Zhenyu  
•
Liu, Fanghui  
•
Chrysos, Grigorios  
Show more
2022
[Proceedings of NeurIPS 2022]
36th Conference on Neural Information Processing Systems - NeurIPS 2022

Neural Architecture Search (NAS) has fostered the automatic discovery of stateof- the-art neural architectures. Despite the progress achieved with NAS, so far there is little attention to theoretical guarantees on NAS. In this work, we study the generalization properties of NAS under a unifying framework enabling (deep) layer skip connection search and activation function search. To this end, we derive the lower (and upper) bounds of the minimum eigenvalue of the Neural Tangent Kernel (NTK) under the (in)finite-width regime using a certain search space including mixed activation functions, fully connected, and residual neural networks. We use the minimum eigenvalue to establish generalization error bounds of NAS in the stochastic gradient descent training. Importantly, we theoretically and experimentally show how the derived results can guide NAS to select the top-performing architectures, even in the case without training, leading to a trainfree algorithm based on our theory. Accordingly, our numerical validation shed light on the design of computationally efficient methods for NAS. Our analysis is non-trivial due to the coupling of various architectures and activation functions under the unifying framework and has its own interest in providing the lower bound of the minimum eigenvalue of NTK in deep learning theory.

  • Files
  • Details
  • Metrics
Type
conference paper
Author(s)
Zhu, Zhenyu  
Liu, Fanghui  
Chrysos, Grigorios  
Cevher, Volkan  orcid-logo
Date Issued

2022

Published in
[Proceedings of NeurIPS 2022]
Total of pages

42

Subjects

ml-ai

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
LIONS  
Event nameEvent placeEvent date
36th Conference on Neural Information Processing Systems - NeurIPS 2022

New Orleans, USA

November 28 - December 3, 2022

Available on Infoscience
October 16, 2022
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/191447
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés