Temporal Segmentation of Facial Behavior

Temporal segmentation of facial gestures in spontaneous facial behavior recorded in real-world settings is an important, unsolved, and relatively unexplored problem in facial image analysis. Several issues contribute to the challenge of this task. These include non-frontal pose, moderate to large out-of-plane head motion, large variability in the temporal scale of facial gestures, and the exponential nature of possible facial action combinations. To address these challenges, we propose a two-step approach to temporally segment facial behavior. The first step uses spectral graph techniques to cluster shape and appearance features invariant to some geometric transformations. The second step groups the clusters into temporally coherent facial gestures. We evaluated this method in facial behavior recorded during face-to- face interactions. The video data were originally collected to answer substantive questions in psychology without concern for algorithm development. The method achieved moderate convergent validity with manual FACS (Facial Action Coding System) annotation. Further, when used to preprocess video for manual FACS annotation, the method significantly improves productivity, thus addressing the need for ground-truth data for facial image analysis. Moreover, we were also able to detect unusual facial behavior.

[1]  P. Ekman,et al.  The ability to detect deceit generalizes across different types of high-stake lies. , 1997, Journal of personality and social psychology.

[2]  Maja Pantic,et al.  Dynamics of facial expression: recognition of facial actions and their temporal segments from face profile image sequences , 2006, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[3]  Nicu Sebe,et al.  Affective multimodal human-computer interaction , 2005, ACM Multimedia.

[4]  W. Rinn,et al.  The neuropsychology of facial expression: a review of the neurological and psychological mechanisms for producing facial expressions. , 1984, Psychological bulletin.

[5]  Aleix M. Martinez Matching expression variant faces , 2003, Vision Research.

[6]  J. Cohn,et al.  Deciphering the Enigmatic Face , 2005, Psychological science.

[7]  Takeo Kanade,et al.  Discriminative cluster analysis , 2006, ICML.

[8]  Jeffrey F. Cohn,et al.  Observer-based measurement of facial expression with the Facial Action Coding System. , 2007 .

[9]  Gwen Littlewort,et al.  Analysis of Machine Learning Methods for Real-Time Recognition of Facial Expressions from Video , 2003 .

[10]  Yajie Tian,et al.  Handbook of face recognition , 2003 .

[11]  Michael J. Black,et al.  Recognizing Facial Expressions in Image Sequences Using Local Parameterized Models of Image Motion , 1997, International Journal of Computer Vision.

[12]  Jordi Vitrià,et al.  Eigenfiltering for flexible eigentracking (EFE) , 2000, Proceedings 15th International Conference on Pattern Recognition. ICPR-2000.

[13]  Ahmed M. Elgammal,et al.  Facial Expression Analysis Using Nonlinear Decomposable Generative Models , 2005, AMFG.

[14]  Simon Baker,et al.  Active Appearance Models Revisited , 2004, International Journal of Computer Vision.

[15]  Lihi Zelnik-Manor,et al.  Temporal Factorization vs. Spatial Factorization , 2004, ECCV.

[16]  J. MacQueen Some methods for classification and analysis of multivariate observations , 1967 .

[17]  Larry S. Davis,et al.  A probabilistic framework for rigid and non-rigid appearance based tracking and recognition , 2000, Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580).

[18]  John J. B. Allen,et al.  The handbook of emotion elicitation and assessment , 2007 .

[19]  Rogério Schmidt Feris,et al.  Manifold Based Analysis of Facial Expression , 2004, 2004 Conference on Computer Vision and Pattern Recognition Workshop.

[20]  P. Ekman,et al.  What the face reveals : basic and applied studies of spontaneous expression using the facial action coding system (FACS) , 2005 .

[21]  Gilad Adiv,et al.  Determining Three-Dimensional Motion and Structure from Optical Flow Generated by Several Moving Objects , 1985, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[22]  Jesse Hoey,et al.  Hierarchical unsupervised learning of facial expression categories , 2001, Proceedings IEEE Workshop on Detection and Recognition of Events in Video.

[23]  P. Ekman,et al.  Facial action coding system: a technique for the measurement of facial movement , 1978 .

[24]  Yiannis Aloimonos,et al.  Deformation and Viewpoint Invariant Color Histograms , 2006, BMVC.

[25]  Anil K. Jain,et al.  Algorithms for Clustering Data , 1988 .

[26]  Jeffrey F. Cohn,et al.  The Timing of Facial Motion in posed and Spontaneous Smiles , 2003, Int. J. Wavelets Multiresolution Inf. Process..

[27]  Rama Chellappa,et al.  Face Processing: Advanced Modeling and Methods , 2006, J. Electronic Imaging.

[28]  Jitendra Malik,et al.  Normalized cuts and image segmentation , 1997, Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[29]  Changbo Hu,et al.  AAM derived face representations for robust facial action recognition , 2006, 7th International Conference on Automatic Face and Gesture Recognition (FGR06).

[30]  Chris H. Q. Ding,et al.  Spectral Relaxation for K-means Clustering , 2001, NIPS.

[31]  Takeo Kanade,et al.  Facial Expression Analysis , 2011, AMFG.

[32]  Chris H. Q. Ding,et al.  K-means clustering via principal component analysis , 2004, ICML.

[33]  W. Rinn Neuropsychology of facial expression. , 1991 .