Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. Structured Prediction of 3D Human Pose with Deep Neural Networks
 
conference paper

Structured Prediction of 3D Human Pose with Deep Neural Networks

Tekin, Bugra  
•
Katircioglu, Isinsu  
•
Salzmann, Mathieu  
Show more
2016
Proceedings of the British Machine Vision Conference (BMVC)
British Machine Vision Conference (BMVC)

Most recent approaches to monocular 3D pose estimation rely on Deep Learning. They either train a Convolutional Neural Network to directly regress from image to 3D pose, which ignores the dependencies between human joints, or model these dependencies via a max-margin structured learning framework, which involves a high computational cost at inference time. In this paper, we introduce a Deep Learning regression architecture for structured prediction of 3D human pose from monocular images that relies on an overcomplete autoencoder to learn a high-dimensional latent pose representation and account for joint dependencies. We demonstrate that our approach outperforms state-of-the-art ones both in terms of structure preservation and prediction accuracy.

  • Files
  • Details
  • Metrics
Loading...
Thumbnail Image
Name

tekin_bmvc16.pdf

Access type

openaccess

Size

799.77 KB

Format

Adobe PDF

Checksum (MD5)

df6a93d51a97373c05f87fc86dc87f11

Loading...
Thumbnail Image
Name

tekin_bmvc16_abstract.pdf

Access type

openaccess

Size

301.62 KB

Format

Adobe PDF

Checksum (MD5)

c6ee7e0a5620a931c948b88b70172fcb

Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés