Scan patterns during the processing of facial expression versus identity: an exploration of task-driven and stimulus-driven effects.

Perceptual studies suggest that processing facial identity emphasizes upper-face information, whereas processing expressions of anger or happiness emphasizes the lower-face. The two goals of the present study were to determine (a) if the distributions of eye fixations reflect these upper/lower-face biases, and (b) whether this bias is task- or stimulus-driven. We presented a target face followed by a probe pair of morphed faces, neither of which was identical to the target. Subjects judged which of the pair was more similar to the target face while eye movements were recorded. In Experiment 1 the probe pair always differed from each other in both identity and expression on each trial. In one block subjects judged which probe face was more similar to the target face in identity, and in a second block subjects judged which probe face was more similar to the target face in expression. In Experiment 2 the two probe faces differed in either expression or identity, but not both. Subjects were not informed which dimension differed, but simply asked to judge which probe face was more similar to the target face. We found that subjects scanned the upper-face more than the lower-face during the identity task but the lower-face more than the upper-face during the expression task in Experiment 1 (task-driven effects), with significantly less variation in bias in Experiment 2 (stimulus-driven effects). We conclude that fixations correlate with regional variations of diagnostic information in different processing tasks, but that these reflect top-down task-driven guidance of information acquisition more than stimulus-driven effects.

[1]  N. J. Cohen,et al.  Eye-movement-based memory effect: a reprocessing effect in face perception. , 1999, Journal of experimental psychology. Learning, memory, and cognition.

[2]  A. Damasio,et al.  The role of scanpaths in facial recognition and learning , 1987, Annals of neurology.

[3]  Alan Agresti,et al.  Categorical Data Analysis , 1991, International Encyclopedia of Statistical Science.

[4]  D. Noton,et al.  Eye movements and visual perception. , 1971, Scientific American.

[5]  M. Hayhoe,et al.  In what ways do eye movements contribute to everyday activities? , 2001, Vision Research.

[6]  J. Shepherd Studies of cue saliency , 1981 .

[7]  Antonio Torralba,et al.  Contextual guidance of eye movements and attention in real-world scenes: the role of global features in object search. , 2006, Psychological review.

[8]  T. Langdell,et al.  Recognition of faces: an approach to the study of autism. , 1978, Journal of child psychology and psychiatry, and allied disciplines.

[9]  G. H. Fisher,et al.  Recognizing human faces. , 1975, Applied ergonomics.

[10]  L. Itti,et al.  Modeling the influence of task on attention , 2005, Vision Research.

[11]  Frédéric Gosselin,et al.  Bubbles: a technique to reveal the use of information in recognition tasks , 2001, Vision Research.

[12]  Xin Chen,et al.  Real-world visual search is dominated by top-down guidance , 2006, Vision Research.

[13]  J. Henderson Human gaze control during real-world scene perception , 2003, Trends in Cognitive Sciences.

[14]  R. Baddeley,et al.  The long and the short of it: Spatial statistics at fixation vary with saccade amplitude and task , 2006, Vision Research.

[15]  Linda J. Lanyon,et al.  A Model of Object-Based Attention That Guides Active Visual Search to Behaviourally Relevant Locations , 2004, WAPCV.

[16]  C. Koch,et al.  A saliency-based search mechanism for overt and covert shifts of visual attention , 2000, Vision Research.

[17]  Derrick J. Parkhurst,et al.  Modeling the role of salience in the allocation of overt visual attention , 2002, Vision Research.

[18]  Gregory J. Zelinsky,et al.  Scene context guides eye movements during visual search , 2006, Vision Research.

[19]  Paula C. Stacey,et al.  Face processing and familiarity: evidence from eye-movement data. , 2005, British journal of psychology.

[20]  Jason J S Barton,et al.  Information Processing during Face Recognition: The Effects of Familiarity, Inversion, and Morphing on Scanning Fixations , 2006, Perception.

[21]  D. S. Wooding,et al.  The relationship between the locations of spatial features and those of fixations made during visual examination of briefly presented images. , 1996, Spatial vision.

[22]  I. Rybak,et al.  A model of attention-guided visual perception and recognition , 1998, Vision Research.

[23]  R. Groner,et al.  Looking at Faces: Local and Global Aspects of Scanpaths , 1984 .

[24]  P. Schyns,et al.  Show Me the Features! Understanding Recognition From the Use of Visual Information , 2002, Psychological science.

[25]  Garrison W. Cottrell,et al.  Transmitting and Decoding Facial Expressions , 2005, Psychological science.

[26]  S M Luria,et al.  Comparison of Eye Movements over Faces in Photographic Positives and Negatives , 1978, Perception.

[27]  D. Barr Analyzing ‘visual world’ eyetracking data using multilevel logistic regression , 2008 .

[28]  G. Deco,et al.  A Neuro-Cognitive Visual System for Object Recognition Based on Testing of Interactive Attentional Top – down Hypotheses , 2000, Perception.

[29]  A. L. I︠A︡rbus Eye Movements and Vision , 1967 .