Perceptually Valid Dynamics for Smiles and Blinks

In many applications, such as conversational agents, virtual reality, movies, and games, animated facial expressions of computer-generated (CG) characters are used to communicate, teach, or entertain. With an increased demand for CG characters, it is important to animate accurate, realistic facial expressions because human facial expressions communicate a wealth of information. However, realistically animating faces is challenging and time-consuming for two reasons. First, human observers are adept at detecting anomalies in realistic CG facial animations. Second, traditional animation techniques based on keyframing sometimes approximate the dynamics of facial expressions or require extensive artistic input while high-resolution performance capture techniques are cost prohibitive. In this thesis, we develop a framework to explore representations of two key facial expressions, blinks and smiles, and we show that data-driven models are needed to realistically animate these expressions. Our approach relies on utilizing high-resolution performance capture data to build models that can be used in traditional keyframing systems. First, we record large collections of high-resolution dynamic expressions through video and motion capture technology. Next, we build expression-specific models of the dynamic data properties of blinks and smiles. We explore variants of the model and assess whether viewers perceive the models as more natural than the simplified models present in the literature. In the first part of the thesis, we build a generative model of the characteristic dynamics of blinks: fast closing of the eyelids followed by a slow opening. Blinks have a characteristic profile with relatively little variation across instances or people. Our results demonstrate the need for an accurate model of eye blink dynamics rather than simple approximations, as viewers perceive the difference. In the second part of the thesis, we investigate how spatial and temporal linearities impact smile genuineness and build a model for genuine smiles. Our perceptual results indicate that a smile model needs to preserve temporal information. With this model, we synthesize perceptually genuine smiles that outperform traditional animation methods accompanied by plausible head motions. In the last part of the thesis, we investigate how blinks synchronize with the start and end of spontaneous smiles. Our analysis shows that eye blinks correlate with the end of the smile and occur before the lip corners stop moving downwards. We argue that the timing of blinks relative to smiles is useful in creating compelling facial expressions. Our work is directly applicable to current methods in animation. For example, we illustrate how our models can be used in the popular framework of blendshape animation to increase realism while keeping the system complexity low. Furthermore, our perceptual results can inform the design of realistic animation systems by highlighting common assumptions that over-simplify the dynamics of expressions.

[1]  Nadia Magnenat-Thalmann,et al.  A computational skin model: fold and wrinkle formation , 2002, IEEE Transactions on Information Technology in Biomedicine.

[2]  P. Ekman,et al.  The Duchenne smile: emotional expression and brain physiology. II. , 1990, Journal of personality and social psychology.

[3]  Heinrich H. Bülthoff,et al.  Render me real? , 2012, ACM Trans. Graph..

[4]  David J. Kalwick Animating facial features & expressions , 2006 .

[5]  N. Kanwisher,et al.  The Fusiform Face Area: A Module in Human Extrastriate Cortex Specialized for Face Perception , 1997, The Journal of Neuroscience.

[6]  M. Bradley,et al.  Affective reactions to briefly presented pictures. , 2001, Psychophysiology.

[7]  Karl F. MacDorman,et al.  Too real for comfort? Uncanny responses to computer generated faces , 2009, Comput. Hum. Behav..

[8]  Jing Xiao,et al.  Multimodal coordination of facial action, head rotation, and eye motion during spontaneous smiles , 2004, Sixth IEEE International Conference on Automatic Face and Gesture Recognition, 2004. Proceedings..

[9]  M. Doughty Consideration of Three Types of Spontaneous Eyeblink Activity in Normal Humans: during Reading and Video Display Terminal Use, in Primary Gaze, and while in Conversation , 2001, Optometry and vision science : official publication of the American Academy of Optometry.

[10]  P. Ekman,et al.  Facial action coding system: a technique for the measurement of facial movement , 1978 .

[11]  Adrian Hilton,et al.  Speech-driven face synthesis from 3D video , 2004 .

[12]  Victor B. Zordan,et al.  Laughing out loud: control for modeling anatomically inspired laughter using audio , 2008, SIGGRAPH 2008.

[13]  Iain Matthews,et al.  Modeling and animating eye blinks , 2011, TAP.

[14]  Catherine Pelachaud,et al.  Greta: A Simple Facial Animation Engine , 2002 .

[15]  Christopher E. Peters,et al.  Attention-driven eye gaze and blinking for virtual humans , 2003, SIGGRAPH '03.

[16]  Takayuki Kanda,et al.  Footing in human-robot conversations: How robots might shape participant roles using gaze cues , 2009, 2009 4th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[17]  P. Caffier,et al.  Experimental evaluation of eye-blink parameters as a drowsiness measure , 2003, European Journal of Applied Physiology.

[18]  A. Manstead,et al.  Can Duchenne smiles be feigned? New evidence on felt and false smiles. , 2009, Emotion.

[19]  B. Knutson Facial expressions of emotion influence interpersonal trait inferences , 1996 .

[20]  Kristin J. Dana,et al.  Skin Texture Modeling , 2005, International Journal of Computer Vision.

[21]  J. Cohn,et al.  All Smiles are Not Created Equal: Morphology and Timing of Smiles Perceived as Amused, Polite, and Embarrassed/Nervous , 2009, Journal of nonverbal behavior.

[22]  Daniel Thalmann,et al.  Abstract muscle action procedures for human face animation , 1988, The Visual Computer.

[23]  Piero Cosi,et al.  Italian consonantal visemes: relationships between spatial/ temporal articulatory characteristics and coproduced acoustic signal , 1997, AVSP.

[24]  Zhaoqi Wang,et al.  Exploring Non‐Linear Relationship of Blendshape Facial Animation , 2011, Comput. Graph. Forum.

[25]  John P. Lewis,et al.  Automated eye motion using texture synthesis , 2005, IEEE Computer Graphics and Applications.

[26]  John Lasseter,et al.  Principles of traditional animation applied to 3D computer animation , 1987, SIGGRAPH.

[27]  Jessica K. Hodgins,et al.  The saliency of anomalies in animated human characters , 2010, TAP.

[28]  Anne M.R. Agur,et al.  Grant's Atlas of Anatomy , 1943 .

[29]  Clifford Nass,et al.  Consistency of personality in interactive characters: verbal cues, non-verbal cues, and user characteristics , 2000, Int. J. Hum. Comput. Stud..

[30]  Sean Andrist,et al.  Designing effective gaze mechanisms for virtual agents , 2012, CHI.

[31]  Wojciech Matusik,et al.  Multi-scale capture of facial geometry and motion , 2007, ACM Trans. Graph..

[32]  S. Porter,et al.  Reading Between the Lies , 2008, Psychological science.

[33]  A. Vrij,et al.  Blinking During and After Lying , 2008 .

[34]  Jessica K. Hodgins,et al.  Perceptual effects of damped and exaggerated facial motion in animated characters , 2013, 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG).

[35]  Eva Krumhuber,et al.  Moving Smiles: The Role of Dynamic Components for the Perception of the Genuineness of Smiles , 2005 .

[36]  Tim Weyrich,et al.  A layered, heterogeneous reflectance model for acquiring and rendering human skin , 2008, SIGGRAPH Asia '08.

[37]  W. P. Smotherman,et al.  Spontaneous eye blinking in human infants: a review. , 2004, Developmental psychobiology.

[38]  Takeo Kanade,et al.  Comprehensive database for facial expression analysis , 2000, Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580).

[39]  Daniel Thalmann,et al.  SMILE: A Multilayered Facial Animation System , 1991, Modeling in Computer Graphics.

[40]  M. Aramideh,et al.  Eyelid movements: behavioral studies of blinking in humans under different stimulus conditions. , 2003, Journal of neurophysiology.

[41]  F. I. Parke June,et al.  Computer Generated Animation of Faces , 1972 .

[42]  U. Dimberg,et al.  Unconscious Facial Reactions to Emotional Facial Expressions , 2000, Psychological science.

[43]  Fred Cummins Gaze and blinking in dyadic conversation: A study in coordinated behaviour among individuals , 2012 .

[44]  J. Veltman,et al.  Physiological workload reactions to increasing levels of task difficulty. , 1998, Ergonomics.

[45]  C. Darwin,et al.  The Expression of the Emotions in Man and Animals , 1956 .

[46]  Hao Li,et al.  Example-based facial rigging , 2010, ACM Transactions on Graphics.

[47]  Ursula Hess,et al.  The cues decoders use in attempting to differentiate emotion-elicited and posed facial expressions , 1994 .

[48]  Moshe Mahler,et al.  Dynamic units of visual speech , 2012, SCA '12.

[49]  Cynthia A. Stifter,et al.  The regulation of positive affect: Gaze aversion activity during mother-infant interaction , 1991 .

[50]  Glenn F. Wilson,et al.  Physiological Data Used to Measure Pilot Workload in Actual Flight and Simulator Conditions , 1987 .

[51]  J. Cohn,et al.  Movement Differences between Deliberate and Spontaneous Facial Expressions: Zygomaticus Major Action in Smiling , 2006, Journal of nonverbal behavior.

[52]  Shigeru Kitazawa,et al.  Eyeblink entrainment at breakpoints of speech , 2010, Neuroscience Research.

[53]  Simon Baker,et al.  Active Appearance Models Revisited , 2004, International Journal of Computer Vision.

[54]  Carlos Busso,et al.  Generating Human-Like Behaviors Using Joint, Speech-Driven Models for Conversational Agents , 2012, IEEE Transactions on Audio, Speech, and Language Processing.

[55]  Jessica K. Hodgins,et al.  Spatial and Temporal Linearities in Posed and Spontaneous Smiles , 2014, ACM Trans. Appl. Percept..

[56]  Peter M. Corcoran,et al.  Statistical models of appearance for eye tracking and eye-blink detection and measurement , 2008, IEEE Transactions on Consumer Electronics.

[57]  Narendra M. Patel,et al.  Parametric Facial Expression Synthesis and Animation , 2010 .

[58]  K. Mardia,et al.  Statistical Shape Analysis , 1998 .

[59]  Verónica Orvalho,et al.  A practical appearance model for dynamic facial color , 2010, SIGGRAPH 2010.

[60]  Albert Ali Salah,et al.  Eyes do not lie: spontaneous versus posed smiles , 2010, ACM Multimedia.

[61]  J. Lochman,et al.  Nonverbal communication of couples in conflict. , 1981 .

[62]  Jeffrey F. Cohn,et al.  The Timing of Facial Motion in posed and Spontaneous Smiles , 2003, Int. J. Wavelets Multiresolution Inf. Process..

[63]  Eva Krumhuber,et al.  Perception of linear and nonlinear motion properties using a FACS validated 3D facial model , 2010, APGV '10.

[64]  Simon Baker,et al.  Lucas-Kanade 20 Years On: A Unifying Framework , 2004, International Journal of Computer Vision.

[65]  Takeo Kanade,et al.  The Extended Cohn-Kanade Dataset (CK+): A complete dataset for action unit and emotion-specified expression , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition - Workshops.

[66]  Yaser Sheikh,et al.  Bilinear spatiotemporal basis models , 2012, TOGS.

[67]  Anthony Steed,et al.  Eyelid kinematics for virtual characters , 2010 .

[68]  Fernando De la Torre,et al.  Interactive region-based linear 3D face models , 2011, SIGGRAPH 2011.

[69]  C. Evinger,et al.  Blinking and associated eye movements in humans, guinea pigs, and rabbits. , 1984, Journal of neurophysiology.

[70]  L. Fleischer Telling Lies Clues To Deceit In The Marketplace Politics And Marriage , 2016 .

[71]  Timothy F. Cootes,et al.  Active Appearance Models , 2001, IEEE Trans. Pattern Anal. Mach. Intell..

[72]  Preliminary Validation of a Multidimensional Model of Wittiness , 1993 .

[73]  Richard Williams,et al.  The Animator's Survival Kit , 2001 .

[74]  Shamus Culhane Animation: From Script to Screen , 1989 .

[75]  E. Ponder,et al.  ON THE ACT OF BLINKING , 1927 .

[76]  Tom Geller,et al.  Overcoming the Uncanny Valley , 2008, IEEE Computer Graphics and Applications.

[77]  Yanxi Liu,et al.  The role of structural facial asymmetry in asymmetry of peak facial expressions , 2006, Laterality.

[78]  Y. Tanaka,et al.  Blink Activity and Task Difficulty , 1993, Perceptual and motor skills.

[79]  George Maestri,et al.  Digital character animation , 1996 .

[80]  Karl F. MacDorman,et al.  Revisiting the uncanny valley theory: Developing and validating an alternative to the Godspeed indices , 2010, Comput. Hum. Behav..

[81]  Dick F. Swaab,et al.  Sex differences in the effects of visual contact and eye contact in negotiations , 2009 .

[82]  Heinrich H. Bülthoff,et al.  Evaluating the perceptual realism of animated facial expressions , 2008, TAP.

[83]  Jun'ichiro Seyama,et al.  The Uncanny Valley: Effect of Realism on the Impression of Artificial Human Faces , 2007, PRESENCE: Teleoperators and Virtual Environments.

[84]  Jessica K. Hodgins,et al.  The temporal connection between smiles and blinks , 2013, 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG).

[85]  W. P. Blount STUDIES OF THE MOVEMENTS OF THE EYELIDS OF ANIMALS: BLINKING , 1927 .

[86]  C. Evinger,et al.  Eyelid movements. Mechanisms and normal data. , 1991, Investigative ophthalmology & visual science.

[87]  Radoslaw Niewiadomski,et al.  Smiling virtual agent in social context , 2012, Cognitive Processing.

[88]  Zhigang Deng,et al.  Rigid Head Motion in Expressive Speech Animation: Analysis and Synthesis , 2007, IEEE Transactions on Audio, Speech, and Language Processing.

[89]  D Guitton,et al.  Upper eyelid movements measured with a search coil during blinks and vertical saccades. , 1991, Investigative ophthalmology & visual science.

[90]  Soraia Raupp Musse,et al.  Evaluation of the Uncanny Valley in CG Characters , 2012, IVA.