Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. SynDeMo: Synergistic Deep Feature Alignment for Joint Learning of Depth and Ego-Motion
 
conference paper

SynDeMo: Synergistic Deep Feature Alignment for Joint Learning of Depth and Ego-Motion

Bozorgtabar, Behzad  
•
Rad, Mohammad Saeed  
•
Mahapatra, Dwarikanath
Show more
January 1, 2019
2019 Ieee/Cvf International Conference On Computer Vision (Iccv 2019)
IEEE/CVF International Conference on Computer Vision (ICCV)

Despite well-established baselines, learning of scene depth and ego-motion from monocular video remains an ongoing challenge, specifically when handling scaling ambiguity issues and depth inconsistencies in image sequences. Much prior work uses either a supervised mode of learning or stereo images. The former is limited by the amount of labeled data, as it requires expensive sensors, while the latter is not always readily available as monocular sequences. In this work, we demonstrate the benefit of using geometric information from synthetic images, coupled with scene depth information, to recover the scale in depth and ego-motion estimation from monocular videos. We developed our framework using synthetic image-depth pairs and unlabeled real monocular images. We had three training objectives: first, to use deep feature alignment to reduce the domain gap between synthetic and monocular images to yield more accurate depth estimation when presented with only real monocular images at test time. Second, we learn scene specific representation by exploiting self-supervision coming from multi-view synthetic images without the need for depth labels. Third, our method uses single-view depth and pose networks, which are capable of jointly training and supervising one another mutually, yielding consistent depth and ego-motion estimates. Extensive experiments demonstrate that our depth and ego-motion models surpass the state-of-the-art, unsupervised methods and compare favorably to early supervised deep models for geometric understanding. We validate the effectiveness of our training objectives against standard benchmarks thorough an ablation study.

  • Details
  • Metrics
Type
conference paper
DOI
10.1109/ICCV.2019.00431
Web of Science ID

WOS:000531438104036

Author(s)
Bozorgtabar, Behzad  
Rad, Mohammad Saeed  
Mahapatra, Dwarikanath
Thiran, Jean-Philippe  
Date Issued

2019-01-01

Publisher

IEEE COMPUTER SOC

Publisher place

Los Alamitos

Published in
2019 Ieee/Cvf International Conference On Computer Vision (Iccv 2019)
ISBN of the book

978-1-7281-4803-8

Series title/Series vol.

IEEE International Conference on Computer Vision

Start page

4209

End page

4218

Subjects

vision

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
LTS5  
Event nameEvent placeEvent date
IEEE/CVF International Conference on Computer Vision (ICCV)

Seoul, SOUTH KOREA

Oct 27-Nov 02, 2019

Available on Infoscience
August 6, 2020
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/170632
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés