Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. Sharp asymptotics on the compression of two-layer neural networks
 
conference paper

Sharp asymptotics on the compression of two-layer neural networks

Amani, Mohammad Hossein  
•
Bombari, Simone
•
Mondelli, Marco
Show more
January 1, 2022
2022 Ieee Information Theory Workshop (Itw)
IEEE Information Theory Workshop (ITW)

In this paper, we study the compression of a target two-layer neural network with N nodes into a compressed network with M < N nodes. More precisely, we consider the setting in which the weights of the target network are i.i.d. sub-Gaussian, and we minimize the population L-2 loss between the outputs of the target and of the compressed network, under the assumption of Gaussian inputs. By using tools from high-dimensional probability, we show that this non-convex problem can be simplified when the target network is sufficiently over-parameterized, and provide the error rate of this approximation as a function of the input dimension and N. In this mean-field limit, the simplified objective, as well as the optimal weights of the compressed network, does not depend on the realization of the target network, but only on expected scaling factors. Furthermore, for networks with ReLU activation, we conjecture that the optimum of the simplified optimization problem is achieved by taking weights on the Equiangular Tight Frame (ETF), while the scaling of the weights and the orientation of the ETF depend on the parameters of the target network. Numerical evidence is provided to support this conjecture.

  • Details
  • Metrics
Type
conference paper
DOI
10.1109/ITW54588.2022.9965870
Web of Science ID

WOS:000904341100099

Author(s)
Amani, Mohammad Hossein  
Bombari, Simone
Mondelli, Marco
Pukdee, Rattana
Rini, Stefano
Date Issued

2022-01-01

Publisher

IEEE

Publisher place

New York

Published in
2022 Ieee Information Theory Workshop (Itw)
ISBN of the book

978-1-6654-8341-4

Series title/Series vol.

Information Theory Workshop

Start page

588

End page

593

Subjects

Computer Science, Information Systems

•

Computer Science, Theory & Methods

•

Mathematics, Applied

•

Computer Science

•

Mathematics

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

Event nameEvent placeEvent date
IEEE Information Theory Workshop (ITW)

Mumbai, INDIA

Nov 01-09, 2022

Available on Infoscience
February 13, 2023
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/194799
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés