Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. GestureGAN for Hand Gesture-to-Gesture Translation in the Wild
 
conference paper

GestureGAN for Hand Gesture-to-Gesture Translation in the Wild

Tang, Hao
•
Wang, Wei  
•
Xu, Dan
Show more
January 1, 2018
Proceedings Of The 2018 Acm Multimedia Conference (Mm'18)
26th ACM Multimedia Conference (MM)

Hand gesture-to-gesture translation in the wild is a challenging task since hand gestures can have arbitrary poses, sizes, locations and self-occlusions. Therefore, this task requires a high-level understanding of the mapping between the input source gesture and the output target gesture. To tackle this problem, we propose a novel hand Gesture Generative Adversarial Network (GestureGAN). GestureGAN consists of a single generator G and a discriminator D, which takes as input a conditional hand image and a target hand skeleton image. GestureGAN utilizes the hand skeleton information explicitly, and learns the gesture-to-gesture mapping through two novel losses, the color loss and the cycle-consistency loss. The proposed color loss handles the issue of "channel pollution" while back-propagating the gradients. In addition, we present the Frechet ResNet Distance (FRD) to evaluate the quality of generated images. Extensive experiments on two widely used benchmark datasets demonstrate that the proposed GestureGAN achieves state-of-the-art performance on the unconstrained hand gesture-to-gesture translation task. Meanwhile, the generated images are in high-quality and are photo-realistic, allowing them to be used as data augmentation to improve the performance of a hand gesture classifier. Our model and code are available at https://github.com/Ha0Tang/GestureGAN.

  • Details
  • Metrics
Type
conference paper
DOI
10.1145/3240508.3240704
Web of Science ID

WOS:000509665700088

Author(s)
Tang, Hao
Wang, Wei  
Xu, Dan
Yan, Yan
Sebe, Nicu
Date Issued

2018-01-01

Publisher

ASSOC COMPUTING MACHINERY

Publisher place

New York

Published in
Proceedings Of The 2018 Acm Multimedia Conference (Mm'18)
ISBN of the book

978-1-4503-5665-7

Start page

774

End page

782

Subjects

Computer Science, Theory & Methods

•

Engineering, Electrical & Electronic

•

Computer Science

•

Engineering

•

generative adversarial networks

•

image translation

•

hand gesture

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
IINFCOM  
Event nameEvent placeEvent date
26th ACM Multimedia Conference (MM)

Seoul, SOUTH KOREA

Oct 22-26, 2018

Available on Infoscience
February 12, 2020
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/165261
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés