Can Your Eyes Tell Me How You Think? A Gaze Directed Estimation of the Mental Activity

We investigate the possibility of estimating the cognitive process used by a person when addressing a mental challenge by following the Eye Accessing Cue (EAC) model from the Neuro-Linguistic Programming (NLP) theory [1]. This model, showed in figure 1, describes the eyemovements that are not used for visual tasks (non visual movements) and suggests that the direction of gaze, in such a case, can be an indicator for the internal representational system used by a person facing a given query. The actual EAC is thought to be identified by distinguishing between the relative position of the iris and the eye socket (lid edge). Our approach is to determine the four limits of the eye socket: the inner and outer corners, the upper and lower lids and the iris center and to subsequently analyze the identified region. The entire method flowchart is presented in figure 2. The schematics of the method used for independently looking for the position of each eye landmark is described in figure 3. Given the face square position by Viola-Jones algorithm [4] and the eye centers given by the method from [3], we fuse information related to position, normalized luminance, template matching and shape constraining. For position and luminance, we construct priors over the training database, while for template matching we describe a patch by concatenation of integral and edge projections on horizontal and vertical directions. The score of how likely is a patch to be centered on the true landmark position is given by a Multi Layer Perceptron. For the shape constrain, inspired by the CLM [2], we construct the probability density function in the eigenspace of the shapes in the training set. By ordering the landmarks according to a prior confidence (e.g. eye outer corners are more reliable than upper and lower eye boundaries) and by keeping all points fixed with the exception of the current least reliable, we build the likelihood of various current landmark positions. This information is fused with previous stages and we iteratively improve the landmark position. The final landmark position is taken as the weighted center of mass of the convex combination between initial stages and shape likelihood. To study the specific of the gaze direction we introduce Eye-Chimera database, which comprises 1172 frontal face images, grouped according to the 7 gaze directions, with a set of 5 points marked for each eye: the iris center and 4 points delimiting the bounding box. Recognizing individual EACs. The recognition of the EAC case (gaze direction) is done by identifying the position of the iris center inside the eye socket complemented by the information of the interior of the eye delimited shape. The interior of the eye quadrilateral shape is described by the integral projections normalized to 32 samples. For the actual recognition we have trained a random forrest to take as input the EAC feature (landmarks positions and integral features). We consider two types of recognition situations: three cases (looking

[1]  Klaus J. Kirchberg,et al.  Robust Face Detection Using the Hausdorff Distance , 2001, AVBPA.

[2]  Daniel S. Messinger,et al.  Automated classification of gaze direction using spectral regression and support vector machine , 2009, 2009 3rd International Conference on Affective Computing and Intelligent Interaction and Workshops.

[3]  Richard W. Thompson,et al.  Mental imagery as revealed by eye movements and spoken predicates: A test of neurolinguistic programming. , 1985 .

[4]  Deva Ramanan,et al.  Face detection, pose estimation, and landmark localization in the wild , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[5]  Paul A. Viola,et al.  Robust Real-Time Face Detection , 2001, Proceedings Eighth IEEE International Conference on Computer Vision. ICCV 2001.

[6]  Mohan M. Trivedi,et al.  Head Pose Estimation in Computer Vision: A Survey , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[7]  Lior Wolf,et al.  An eye for an eye: A single camera gaze-replacement method , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[8]  Heiko Neumann,et al.  A comprehensive head pose and gaze database , 2007 .

[9]  Corneliu Florea,et al.  Zero-crossing based image projections encoding for eye localization , 2012, 2012 Proceedings of the 20th European Signal Processing Conference (EUSIPCO).

[10]  Laura Chamberlain Eye Tracking Methodology; Theory and Practice , 2007 .

[11]  Montse Pardàs,et al.  Edge projections for eye localization , 2008 .

[12]  Adam Schmidt,et al.  The put face database , 2008 .

[13]  C. E. Beck,et al.  Test of the Eye-Movement Hypothesis of Neurolinguistic Programming: A Rebuttal of Conclusions , 1984 .

[14]  D I Perrett,et al.  Organization and functions of cells responsive to faces in the temporal cortex. , 1992, Philosophical transactions of the Royal Society of London. Series B, Biological sciences.

[15]  Stefanos Kollias,et al.  A natural head pose and eye gaze dataset , 2009, AFFINE '09.

[16]  Theo Gevers,et al.  Accurate eye center location and tracking using isophote curvature , 2008, 2008 IEEE Conference on Computer Vision and Pattern Recognition.

[17]  R. Kleck,et al.  Effects of direct and averted gaze on the perception of facially communicated emotion. , 2005, Emotion.

[18]  Qiang Ji,et al.  Automatic Eye Detection and Its Validation , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05) - Workshops.

[19]  Simon Lucey,et al.  Deformable Model Fitting by Regularized Landmark Mean-Shift , 2010, International Journal of Computer Vision.

[20]  Bülent Sankur,et al.  A comparative study of face landmarking techniques , 2013, EURASIP J. Image Video Process..

[21]  Wendy Robertson,et al.  Neurolinguistic programming: a systematic review of the effects on health outcomes. , 2012, The British journal of general practice : the journal of the Royal College of General Practitioners.

[22]  M. Buckner,et al.  Eye movement as an indicator of sensory components in thought. , 1987 .

[23]  David J. Kriegman,et al.  Localizing parts of faces using a consensus of exemplars , 2011, CVPR.

[24]  Qiang Ji,et al.  In the Eye of the Beholder: A Survey of Models for Eyes and Gaze , 2010, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[25]  Timothy F. Cootes,et al.  Feature Detection and Tracking with Constrained Local Models , 2006, BMVC.

[26]  Paul A. Bromiley,et al.  Robust and Accurate Shape Model Matching Using Random Forest Regression-Voting , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[27]  S A Poffel,et al.  Neurolinguistic Programming: A Test of the Eye-Movement Hypothesis , 1985, Perceptual and motor skills.

[28]  Richard Bandler,et al.  Frogs into princes : neuro linguistic programming , 1979 .

[29]  Timothy F. Cootes,et al.  Active Shape Models-Their Training and Application , 1995, Comput. Vis. Image Underst..

[30]  Juliet Grayson,et al.  Neuro-Linguistic Programming , 2000 .

[31]  D. Spalding The Principles of Psychology , 1873, Nature.

[32]  Timothy F. Cootes,et al.  Active Appearance Models , 2001, IEEE Trans. Pattern Anal. Mach. Intell..

[33]  Maja Pantic,et al.  Facial point detection using boosted regression and graph models , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[34]  Theo Gevers,et al.  Accurate Eye Center Location through Invariant Isocentric Patterns , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[35]  G. Underwood Cognitive processes in eye guidance , 2005 .

[36]  T. K. Leungfj,et al.  Finding Faces in Cluttered Scenes using Random Labeled Graph Matching , 1995 .