Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. Video Prediction Recalling Long-term Motion Context via Memory Alignment Learning
 
conference paper

Video Prediction Recalling Long-term Motion Context via Memory Alignment Learning

Lee, Sangmin
•
Kim, Hak Gu  
•
Choi, Dae Hwi
Show more
January 1, 2021
2021 Ieee/Cvf Conference On Computer Vision And Pattern Recognition, Cvpr 2021
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)

Our work addresses long-term motion context issues for predicting future frames. To predict the future precisely, it is required to capture which long-term motion context (e.g., walking or running) the input motion (e.g., leg movement) belongs to. The bottlenecks arising when dealing with the long-term motion context are: (i) how to predict the long-term motion context naturally matching input sequences with limited dynamics, (ii) how to predict the long-term motion context with high-dimensionality (e.g., complex motion). To address the issues, we propose novel motion context-aware video prediction. To solve the bottleneck (i), we introduce a long-term motion context memory (LMC-Memory) with memory alignment learning. The proposed memory alignment learning enables to store long-term motion contexts into the memory and to match them with sequences including limited dynamics. As a result, the long-term context can be recalled from the limited input sequence. In addition, to resolve the bottleneck (ii), we propose memory query decomposition to store local motion context (i.e., low-dimensional dynamics) and recall the suitable local context for each local part of the input individually. It enables to boost the alignment effects of the memory. Experimental results show that the proposed method outperforms other sophisticated RNN-based methods, especially in long-term condition. Further, we validate the effectiveness of the proposed network designs by conducting ablation studies and memory feature analysis. The source code of this work is available(dagger).

  • Details
  • Metrics
Type
conference paper
DOI
10.1109/CVPR46437.2021.00307
Web of Science ID

WOS:000739917303025

Author(s)
Lee, Sangmin
Kim, Hak Gu  
Choi, Dae Hwi
Kim, Hyung-Il
Ro, Yong Man
Date Issued

2021-01-01

Publisher

IEEE COMPUTER SOC

Publisher place

Los Alamitos

Published in
2021 Ieee/Cvf Conference On Computer Vision And Pattern Recognition, Cvpr 2021
ISBN of the book

978-1-6654-4509-2

Series title/Series vol.

IEEE Conference on Computer Vision and Pattern Recognition

Start page

3053

End page

3062

Subjects

Computer Science, Artificial Intelligence

•

Imaging Science & Photographic Technology

•

Computer Science

•

Imaging Science & Photographic Technology

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
IVRL  
Event nameEvent placeEvent date
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)

ELECTR NETWORK

Jun 19-25, 2021

Available on Infoscience
February 14, 2022
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/185365
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés