Online Modeling For Realtime Facial Animation

We present a new algorithm for realtime face tracking on commodity RGB-D sensing devices. Our method requires no user-specific training or calibration, or any other form of manual assistance, thus enabling a range of new applications in performance-based facial animation and virtual interaction at the consumer level. The key novelty of our approach is an optimization algorithm that jointly solves for a detailed 3D expression model of the user and the corresponding dynamic tracking parameters. Realtime performance and robust computations are facilitated by a novel subspace parameterization of the dynamic facial expression space. We provide a detailed evaluation that shows that our approach significantly simplifies the performance capture workflow, while achieving accurate facial tracking for realtime applications.


Published in:
Acm Transactions On Graphics, 32, 4
Year:
2013
Publisher:
New York, Assoc Computing Machinery
ISSN:
0730-0301
Keywords:
Laboratories:




 Record created 2013-10-01, last modified 2018-03-17


Rate this document:

Rate this document:
1
2
3
 
(Not yet reviewed)