A Deep Learning Approach for Robust Head Pose Independent Eye Movements Recognition from Videos

Recognizing eye movements is important for gaze behavior understanding like in human communication analysis (human-human or robot interactions) or for diagnosis (medical, reading impairments). In this paper, we address this task using remote RGB-D sensors to analyze people behaving in natural conditions. This is very challenging given that such sensors have a normal sampling rate of 30 Hz and provide low-resolution eye images (typically 36x60 pixels), and natural scenarios introduce many variabilities in illumination, shadows, head pose, and dynamics. Hence gaze signals one can extract in these conditions have lower precision compared to dedicated IR eye trackers, rendering previous methods less appropriate for the task. To tackle these challenges, we propose a deep learning method that directly processes the eye image video streams to classify them into fixation, saccade, and blink classes, and allows to distinguish irrelevant noise (illumination, low-resolution artifact, inaccurate eye alignment, difficult eye shapes) from true eye motion signals. Experiments on natural 4-party interactions demonstrate the benefit of our approach compared to previous methods, including deep learning models applied to gaze outputs.


Publié dans:
5
Présenté à:
2019 ACM Symposium on Eye Tracking Research & Applications
Année
2019
Publisher:
ACM
Mots-clefs:
Laboratoires:




 Notice créée le 2019-03-25, modifiée le 2019-03-31

Lien externe:
Télécharger le document
Related documents
Évaluer ce document:

Rate this document:
1
2
3
 
(Pas encore évalué)