A Deep Learning Approach for Robust Head Pose Independent Eye Movements Recognition from Videos

Recognizing eye movements is important for gaze behavior understanding like in human communication analysis (human-human or robot interactions) or for diagnosis (medical, reading impairments). In this paper, we address this task using remote RGB-D sensors to analyze people behaving in natural conditions. This is very challenging given that such sensors have a normal sampling rate of 30 Hz and provide low-resolution eye images (typically 36x60 pixels), and natural scenarios introduce many variabilities in illumination, shadows, head pose, and dynamics. Hence gaze signals one can extract in these conditions have lower precision compared to dedicated IR eye trackers, rendering previous methods less appropriate for the task. To tackle these challenges, we propose a deep learning method that directly processes the eye image video streams to classify them into fixation, saccade, and blink classes, and allows to distinguish irrelevant noise (illumination, low-resolution artifact, inaccurate eye alignment, difficult eye shapes) from true eye motion signals. Experiments on natural 4-party interactions demonstrate the benefit of our approach compared to previous methods, including deep learning models applied to gaze outputs.


Published in:
5
Presented at:
2019 ACM Symposium on Eye Tracking Research & Applications
Year:
2019
Publisher:
ACM
Keywords:
Laboratories:




 Record created 2019-03-25, last modified 2019-08-12

External link:
Download fulltext
Related documents
Rate this document:

Rate this document:
1
2
3
 
(Not yet reviewed)