Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. Side-tuning: a Baseline for Network Adaptation Via Additive Side Networks
 
conference paper

Side-tuning: a Baseline for Network Adaptation Via Additive Side Networks

Zhang, Jeffrey O.
•
Sax, Alexander
•
Zamir, Amir  
Show more
Vedaldi, A
•
Bischof, H
Show more
December 3, 2020
Computer Vision – ECCV 2020 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings
16th European Conference on Computer Vision Workshops (ECCV 2020)

When training a neural network for a desired task, one may prefer to adapt a pre-trained network rather than starting from randomly initialized weights. Adaptation can be useful in cases when training data is scarce, when a single learner needs to perform multiple tasks, or when one wishes to encode priors in the network. The most commonly employed approaches for network adaptation are fine-tuning and using the pre-trained network as a fixed feature extractor, among others. In this paper, we propose a straightforward alternative: side-tuning. Side-tuning adapts a pre-trained network by training a lightweight "side" network that is fused with the (unchanged) pre-trained network via summation. This simple method works as well as or better than existing solutions and it resolves some of the basic issues with fine-tuning, fixed features, and other common approaches. In particular, side-tuning is less prone to overfitting, is asymptotically consistent, and does not suffer from catastrophic forgetting in incremental learning. We demonstrate the performance of side-tuning under a diverse set of scenarios, including incremental learning (iCIFAR, iTaskonomy), reinforcement learning, imitation learning (visual navigation in Habitat), NLP question-answering (SQuAD v2), and single-task transfer learning (Taskonomy), with consistently promising results.

  • Details
  • Metrics
Type
conference paper
DOI
10.1007/978-3-030-58580-8_41
Web of Science ID

WOS:001500572400041

Author(s)
Zhang, Jeffrey O.

University of California System

Sax, Alexander

University of California System

Zamir, Amir  

École Polytechnique Fédérale de Lausanne

Guibas, Leonidas

Stanford University

Malik, Jitendra

University of California System

Editors
Vedaldi, A
•
Bischof, H
•
Brox, T
•
Frahm, JM
Date Issued

2020-12-03

Publisher

Springer Nature

Publisher place

Cham

Published in
Computer Vision – ECCV 2020 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings
ISBN of the book

978-3-030-58579-2

978-3-030-58580-8

Book part number

Part III

Series title/Series vol.

Lecture Notes in Computer Science; 12348

ISSN (of the series)

0302-9743

Start page

698

End page

714

Subjects

Sidetuning

•

Finetuning

•

Transfer learning

•

Representation learning

•

Lifelong learning

•

Incremental learning

•

Continual learning

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
VILAB  
Event nameEvent acronymEvent placeEvent date
16th European Conference on Computer Vision Workshops (ECCV 2020)

ECCV 2020

Glasgow, UK

2020-08-23 - 2020-08-28

FunderFunding(s)Grant NumberGrant URL

MURI

N00014-14-1-0671

Vannevar Bush Faculty Fellowship

Amazon AWS Machine Learning Award

Show more
Available on Infoscience
June 25, 2025
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/251547
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés