Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Journal articles
  4. Hierarchical Training of Deep Neural Networks Using Early Exiting
 
research article

Hierarchical Training of Deep Neural Networks Using Early Exiting

Sepehri, Yamin  
•
Pad, Pedram
•
Yuzuguler, Ahmet Caner  
Show more
May 14, 2024
Ieee Transactions On Neural Networks And Learning Systems

Deep neural networks (DNNs) provide state-of-the-art accuracy for vision tasks, but they require significant resources for training. Thus, they are trained on cloud servers far from the edge devices that acquire the data. This issue increases communication cost, runtime, and privacy concerns. In this study, a novel hierarchical training method for DNNs is proposed that uses early exits in a divided architecture between edge and cloud workers to reduce the communication cost, training runtime, and privacy concerns. The method proposes a brand-new use case for early exits to separate the backward pass of neural networks between the edge and the cloud during the training phase. We address the issues of most available methods that, due to the sequential nature of the training phase, cannot train the levels of hierarchy simultaneously or they do it with the cost of compromising privacy. In contrast, our method can use both edge and cloud workers simultaneously, does not share the raw input data with the cloud, and does not require communication during the backward pass. Several simulations and on-device experiments for different neural network architectures demonstrate the effectiveness of this method. It is shown that the proposed method reduces the training runtime for VGG-16 and ResNet-18 architectures by 29% and 61% in CIFAR-10 classification and by 25% and 81% in Tiny ImageNet classification, respectively, when the communication with the cloud is done over a low bit rate channel. This gain in the runtime is achieved, while the accuracy drop is negligible. This method is advantageous for online learning of high-accuracy DNNs on sensor-holding low-resource devices such as mobile phones or robots as a part of an edge-cloud system, making them more flexible in facing new tasks and classes of data.

  • Details
  • Metrics
Type
research article
DOI
10.1109/TNNLS.2024.3396628
Web of Science ID

WOS:001226168300001

Author(s)
Sepehri, Yamin  
Pad, Pedram
Yuzuguler, Ahmet Caner  
Frossard, Pascal  
Dunbar, Andrea  

EPFL

Date Issued

2024-05-14

Publisher

Ieee-Inst Electrical Electronics Engineers Inc

Published in
Ieee Transactions On Neural Networks And Learning Systems
Subjects

Technology

•

Training

•

Costs

•

Runtime

•

Cloud Computing

•

Task Analysis

•

Artificial Neural Networks

•

Privacy

•

Deep Learning

•

Early Exiting

•

Edge-Cloud Systems

•

Hierarchical Training

•

Neural Network

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
LTS4  
MTEI-GE  
Available on Infoscience
June 19, 2024
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/208613
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés