Accurate markerless jaw tracking for facial performance capture

We present the first method to accurately track the invisible jaw based solely on the visible skin surface, without the need for any markers or augmentation of the actor. As such, the method can readily be integrated with off-the-shelf facial performance capture systems. The core idea is to learn a non-linear mapping from the skin deformation to the underlying jaw motion on a dataset where ground-truth jaw poses have been acquired, and then to retarget the mapping to new subjects. Solving for the jaw pose plays a central role in visual effects pipelines, since accurate jaw motion is required when retargeting to fantasy characters and for physical simulation. Currently, this task is performed mostly manually to achieve the desired level of accuracy, and the presented method has the potential to fully automate this labour intense and error prone process.

[1]  G. Throckmorton,et al.  Reducing within-subject variation in chewing cycle kinematics--a statistical approach. , 2004, Archives of oral biology.

[2]  Justus Thies,et al.  Face2Face: real-time face capture and reenactment of RGB videos , 2019, Commun. ACM.

[3]  E Nordh,et al.  Wireless optoelectronic recordings of mandibular and associated head-neck movements in man: a methodological study. , 2000, Journal of oral rehabilitation.

[4]  Srikumar Ramalingam,et al.  Building anatomically realistic jaw kinematics model from data , 2018, The Visual Computer.

[5]  Derek Bradley,et al.  Rigid stabilization of facial expressions , 2014, ACM Trans. Graph..

[6]  Andrew Jones,et al.  Multi‐View Stereo on Consistent Face Topology , 2017, Comput. Graph. Forum.

[7]  Erin M Wilson,et al.  A kinematic description of the temporal characteristics of jaw motion for early chewing: preliminary findings. , 2012, Journal of speech, language, and hearing research : JSLHR.

[8]  Wan-Chun Ma,et al.  Comprehensive Facial Performance Capture , 2011, Comput. Graph. Forum.

[9]  Weiliang Xu,et al.  A Novel Spatial Mandibular Motion-Capture System Based on Planar Fiducial Markers , 2018, IEEE Sensors Journal.

[10]  Jihun Yu,et al.  Realtime facial animation with on-the-fly correctives , 2013, ACM Trans. Graph..

[11]  Ira Kemelmacher-Shlizerman,et al.  Total Moving Face Reconstruction , 2014, ECCV.

[12]  Kun Zhou,et al.  Displaced dynamic expression regression for real-time facial tracking and animation , 2014, ACM Trans. Graph..

[13]  P. Bracco,et al.  Deep bite: a case report with chewing pattern and electromyographic activity before and after therapy with function generating bite. , 2013, European Journal of Paediatric Dentistry.

[14]  Jihun Yu,et al.  Unconstrained realtime facial performance capture , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[15]  Taiji Sohmura,et al.  Measurement of masticatory movement by a new jaw tracking system using a home digital camcorder. , 2005, Dental materials journal.

[16]  Christian Theobalt,et al.  Reconstructing detailed dynamic face geometry from monocular video , 2013, ACM Trans. Graph..

[17]  Jaakko Lehtinen,et al.  Production-level facial performance capture using deep convolutional neural networks , 2016, Symposium on Computer Animation.

[18]  Derek Bradley,et al.  An empirical rig for jaw animation , 2018, ACM Trans. Graph..

[19]  E Nordh,et al.  Co-ordinated Mandibular and Head-Neck Movements during Rhythmic Jaw Activities in Man , 2000, Journal of dental research.

[20]  Thabo Beeler,et al.  Real-time high-fidelity facial performance capture , 2015, ACM Trans. Graph..

[21]  A. Andrade,et al.  A COMPUTATIONAL METHOD FOR RECORDING AND ANALYSIS OF MANDIBULAR MOVEMENTS , 2008, Journal of applied oral science : revista FOB.

[22]  K. Nishigawa,et al.  Current status of researches on jaw movement and occlusion for clinical application , 2009 .

[23]  Xin Tong,et al.  Automatic acquisition of high-fidelity facial performances using monocular videos , 2014, ACM Trans. Graph..

[24]  Hao Li,et al.  Realtime performance-based facial animation , 2011, ACM Trans. Graph..

[25]  Justus Thies,et al.  Headon , 2018, ACM Trans. Graph..

[26]  G. Throckmorton,et al.  Chewing cycle kinematics of subjects with deepbite malocclusion. , 2007, American journal of orthodontics and dentofacial orthopedics : official publication of the American Association of Orthodontists, its constituent societies, and the American Board of Orthodontics.

[27]  V. Ferrario,et al.  Quantification of translational and gliding components in human temporomandibular joint during mouth opening. , 2005, Archives of oral biology.

[28]  Yoshinobu Maeda,et al.  Markerless three-dimensional tracking of masticatory movement. , 2016, Journal of biomechanics.

[29]  Jaw-opening accuracy is not affected by masseter muscle vibration in healthy men , 2014, Experimental Brain Research.

[30]  J. Gower Generalized procrustes analysis , 1975 .

[31]  João Manuel R. S. Tavares,et al.  A system for analysis of the 3D mandibular movement using magnetic sensors and neuronal networks , 2006 .

[32]  Oscar Cordón,et al.  Genetic algorithms for skull-face overlay including mandible articulation , 2017, Inf. Sci..

[33]  Yangang Wang,et al.  Online modeling for realtime facial animation , 2013, ACM Trans. Graph..

[34]  Davis E. King,et al.  Dlib-ml: A Machine Learning Toolkit , 2009, J. Mach. Learn. Res..

[35]  W. Heidrich,et al.  High resolution passive facial performance capture , 2010, ACM Trans. Graph..

[36]  H. Hayasaki,et al.  Quantification of human chewing-cycle kinematics. , 2000, Archives of oral biology.

[37]  Patrick Pérez,et al.  MoFA: Model-Based Deep Convolutional Face Autoencoder for Unsupervised Monocular Reconstruction , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[38]  Derek Bradley,et al.  High resolution passive facial performance capture , 2010, SIGGRAPH 2010.

[39]  Justus Thies,et al.  Real-time expression transfer for facial reenactment , 2015, ACM Trans. Graph..

[40]  Hans-Peter Seidel,et al.  Lightweight binocular facial performance capture under uncontrolled lighting , 2012, ACM Trans. Graph..

[41]  Derek Bradley,et al.  An anatomically-constrained local deformation model for monocular face capture , 2016, ACM Trans. Graph..

[42]  Derek Bradley,et al.  High-quality passive facial performance capture using anchor frames , 2011, ACM Trans. Graph..

[43]  Justus Thies,et al.  Demo of Face2Face: real-time face capture and reenactment of RGB videos , 2016, SIGGRAPH Emerging Technologies.

[44]  J F Prinz The Cybermouse: a simple method of describing the trajectory of the human mandible in three dimensions. , 1997, Journal of biomechanics.