Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Journal articles
  4. An Analysis of Super-Net Heuristics in Weight-Sharing NAS
 
research article

An Analysis of Super-Net Heuristics in Weight-Sharing NAS

Yu, Kaicheng  
•
Ranftl, Rene
•
Salzmann, Mathieu  
November 1, 2022
Ieee Transactions On Pattern Analysis And Machine Intelligence

Weight sharing promises to make neural architecture search (NAS) tractable even on commodity hardware. Existing methods in this space rely on a diverse set of heuristics to design and train the shared-weight backbone network, a.k.a. the super-net. Since heuristics substantially vary across different methods and have not been carefully studied, it is unclear to which extent they impact super-net training and hence the weight-sharing NAS algorithms. In this paper, we disentangle super-net training from the search algorithm, isolate 14 frequently-used training heuristics, and evaluate them over three benchmark search spaces. Our analysis uncovers that several commonly-used heuristics negatively impact the correlation between super-net and stand-alone performance, whereas simple, but often overlooked factors, such as proper hyper-parameter settings, are key to achieve strong performance. Equipped with this knowledge, we show that simple random search achieves competitive performance to complex state-of-the-art NAS algorithms when the super-net is properly trained.

  • Details
  • Metrics
Type
research article
DOI
10.1109/TPAMI.2021.3108480
Web of Science ID

WOS:000864325900060

Author(s)
Yu, Kaicheng  
Ranftl, Rene
Salzmann, Mathieu  
Date Issued

2022-11-01

Publisher

IEEE COMPUTER SOC

Published in
Ieee Transactions On Pattern Analysis And Machine Intelligence
Volume

44

Issue

11

Start page

8110

End page

8124

Subjects

Computer Science, Artificial Intelligence

•

Engineering, Electrical & Electronic

•

Computer Science

•

Engineering

•

training

•

protocols

•

computer architecture

•

task analysis

•

measurement

•

benchmark testing

•

encoding

•

automl

•

neural architecture search

•

weight-sharing

•

super-net

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
CVLAB  
Available on Infoscience
November 7, 2022
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/191945
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés