Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Journal articles
  4. High-performance deep spiking neural networks with 0.3 spikes per neuron
 
research article

High-performance deep spiking neural networks with 0.3 spikes per neuron

Stanojevic, Ana  
•
Wozniak, Stanislaw Andrzej  
•
Bellec, Guillaume  
Show more
August 9, 2024
Nature Communications

Communication by rare, binary spikes is a key factor for the energy efficiency of biological brains. However, it is harder to train biologically-inspired spiking neural networks than artificial neural networks. This is puzzling given that theoretical results provide exact mapping algorithms from artificial to spiking neural networks with time-to-first-spike coding. In this paper we analyze in theory and simulation the learning dynamics of time-to-first-spike-networks and identify a specific instance of the vanishing-or-exploding gradient problem. While two choices of spiking neural network mappings solve this problem at initialization, only the one with a constant slope of the neuron membrane potential at threshold guarantees the equivalence of the training trajectory between spiking and artificial neural networks with rectified linear units. For specific image classification architectures comprising feed-forward dense or convolutional layers, we demonstrate that deep spiking neural network models can be effectively trained from scratch on MNIST and Fashion-MNIST datasets, or fine-tuned on large-scale datasets, such as CIFAR10, CIFAR100 and PLACES365, to achieve the exact same performance as that of artificial neural networks, surpassing previous spiking neural networks. Our approach accomplishes high-performance classification with less than 0.3 spikes per neuron, lending itself for an energy-efficient implementation. We also show that fine-tuning spiking neural networks with our robust gradient descent algorithm enables their optimization for hardware implementations with low latency and resilience to noise and quantization.

  • Files
  • Details
  • Metrics
Type
research article
DOI
10.1038/s41467-024-51110-5
Author(s)
Stanojevic, Ana  

IBM Research - Zurich

Wozniak, Stanislaw Andrzej  

IBM Research - Zurich

Bellec, Guillaume  

EPFL

Cherubini, Giovanni

IBM Research - Zurich

Pantazi, Angeliki  
Gerstner, Wulfram  

EPFL

Date Issued

2024-08-09

Publisher

Springer Science and Business Media LLC

Published in
Nature Communications
Volume

15

Issue

1

Article Number

6793

URL

View on Nature.com

https://www.nature.com/articles/s41467-024-51110-5
Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
LCN1  
FunderFunding(s)Grant NumberGrant URL

Swiss National Science Foundation

Sinergia

CRSII5 198612

Available on Infoscience
August 12, 2024
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/240693
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés