Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. Deep Multitask Gaze Estimation with a Constrained Landmark-Gaze Model
 
conference paper

Deep Multitask Gaze Estimation with a Constrained Landmark-Gaze Model

Yu, Yu
•
Liu, Gang
•
Odobez, Jean-Marc  
2018
Computer Vision - Eccv 2018 Workshops, Pt Ii
15th European Conference on Computer Vision (ECCV)

As an indicator of attention, gaze is an important cue for human behavior and social interaction analysis. Recent deep learning methods for gaze estimation rely on plain regression of the gaze from images without accounting for potential mismatches in eye image cropping and normalization. This may impact the estimation of the implicit relation between visual cues and the gaze direction when dealing with low resolution images or when training with a limited amount of data. In this paper, we propose a deep multitask framework for gaze estimation, with the following contributions. (i) we proposed a multitask framework which relies on both synthetic data and real data for end-to-end training. During training, each dataset provides the label of only one task but the two tasks are combined in a constrained way. (ii) we introduce a Constrained Landmark-Gaze Model (CLGM) modeling the joint variation of eye landmark locations (including the iris center) and gaze directions. By relating explicitly visual information (landmarks) to the more abstract gaze values, we demonstrate that the estimator is more accurate and easier to learn. (iii) by decomposing our deep network into a network inferring jointly the parameters of the CLGM model and the scale and translation parameters of eye regions on one hand, and a CLGM based decoder deterministically inferring landmark positions and gaze from these parameters and head pose on the other hand, our framework decouples gaze estimation from irrelevant geometric variations in the eye image (scale, translation), resulting in a more robust model. Thorough experiments on public datasets demonstrate that our method achieves competitive results, improving over state-of-the-art results in challenging free head pose gaze estimation tasks and on eye landmark localization (iris location) ones.

  • Details
  • Metrics
Type
conference paper
DOI
10.1007/978-3-030-11012-3_35
Author(s)
Yu, Yu
Liu, Gang
Odobez, Jean-Marc  
Date Issued

2018

Publisher

SPRINGER INTERNATIONAL PUBLISHING AG

Publisher place

Cham

Published in
Computer Vision - Eccv 2018 Workshops, Pt Ii
ISBN of the book

978-3-030-11011-6

978-3-030-11012-3

Series title/Series vol.

Lecture Notes in Computer Science

Volume

11130

Start page

456

End page

474

URL

Related documents

https://publidiap.idiap.ch/downloads//papers/2018/Yu_ECCVW_2018.pdf

Related documents

https://publidiap.idiap.ch/index.php/publications/showcite/Yu_Idiap-Internal-RR-05-2018
Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
LIDIAP  
Event nameEvent placeEvent date
15th European Conference on Computer Vision (ECCV)

Munich, GERMANY

Sep 08-14, 2018

Available on Infoscience
January 22, 2019
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/153653
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés