Person, place, and past influence eye movements during visual search

Person, place, and past influence eye movements during visual search Barbara Hidalgo-Sotelo (bhs@mit.edu) Aude Oliva (oliva@mit.edu) Department of Brain and Cognitive Sciences Massachusetts Institute of Technology Cambridge, MA 02139 Abstract features directs gaze to expectation-based scene regions (Eckstein, Drescher & Shimozaki, 2006; Henderson, 2003; Torralba, Oliva, Castelhano, & Henderson, 2006). At the level of scene exemplar representations, spatial context can also be used to allocate attention preferentially to regions that have become associated with the target. In contextual cueing, for example, observers implicitly learn patterns in repeated displays that help them find a target faster in repeated configurations (Chun & Jiang, 1998). It is not well understood, however, whether a scene exemplar representation can systematically bias individual fixations. How can “experience based” influences be distinguished from the myriad of sources that guide attention to relevant scene regions? One challenge is that attention is strongly guided by information that does not depend on specific experience. Figure 1 illustrates regularities in eye fixations across and within observers. In Figure 1A, fixations from 9 observers searching for a book are shown; the high fixation density along countertop surfaces illustrates how spatial layout and context guide where observers look. Systematic biases unrelated to the scene’s content also influence gaze location. In Figure 1B, fixations sampled from random scenes have been projected onto the kitchen scene. Center bias in the fixation distribution is driven by oculomotor tendencies (Tatler, 2007; Tatler & Vincent, 2009) and photographer bias. A second challenge, of the opposing nature, lies in the significant variability in fixation locations across individuals. As a result, two independent observers may fixate different scene regions, even when looking for the same object in the same scene (Figure 1C). It is possible that individuals are biased by experience, but that the effects are masked by pooling over experienced observers. Given initial differences in search patterns, could systematic differences arise when an observer repeats her search of the scene? To reasonably estimate the influence of past experience, the search patterns of observers who have never viewed the scene must be contrasted with different observers who have previously searched the scene. In this paper, eye movement data from a visual search study was analyzed using approach we have termed comparative map analysis. This analysis was used to evaluate how different sources of information contribute to attentional guidance during visual search of familiar scenes. In our experiment, observers’ eyes were tracked while they looked for a book in pictures of real world scenes. On some trials, observers searched a scene that had been presented previously. Importantly, the target object and location remained unchanged in each presentation of the scene. The What is the role of an individual’s past experience in guiding gaze in familiar environments? Contemporary models of search guidance suggest high level scene context is a strong predictor of where observers search in realistic scenes. Specific associations also develop between particular places and object locations. Together, scene context and place- specific associations bias attention to informative spatial locations. At the level of eye fixations, it is not known whether a person’s specific search experience influences attentional selection. Eye movements are notoriously variable: people often foveate different places when searching for the same target in the same scene. Do individual differences in fixation locations influence how a scene is subsequently examined? We introduce a method, comparative map analysis, for analyzing spatial patterns in eye movement data. Using this method, we quantified the consistency of fixated locations within the same observer and between observers during search of real world scenes. Results indicated a remarkable consistency in the locations fixated by the same observer across multiple searches of a given scene. This observer-specific guidance was shown to be distinct from general scene context information or familiarity with the scene. Accordingly, this is considered evidence for a uniquely informative role of an individual’s search experience on attentional guidance in a familiar scene. Keywords: visual search; eye movements; scene perception; learning; attentional guidance; fixation similarity Introduction An important feature of ecological visual search is that there are few truly novel, unfamiliar places in which a person is likely to search. Many tasks involve examining the same place repeatedly, such as the various times spent searching for a specific utensil in one’s own kitchen. Locating the target in question benefits from both category based information (e.g. utensils are on countertops) and place specific information (e.g. in this kitchen, utensils hang over the stove). For any observer, there will be many sources of information that guide which scene regions are inspected during search. What influence does a person’s own search experience (i.e. fixation locations) have in guiding where they are likely to look in familiar scenes? A growing body of evidence suggests that observers use high level information, such as learned target features and global scene context, to guide their gaze when searching for an object in real world environments (Ehinger, Hidalgo- Sotelo, Torralba, & Oliva, 2009; Hwang, Higgins, Pomplun, 2009). At this level of categorical representation, knowledge of the basic-level scene category and target

[1]  B. Tatler,et al.  The prominence of behavioural biases in eye guidance , 2009 .

[2]  Krista A. Ehinger,et al.  Modelling search for people in 900 scenes: A combined source model of eye guidance , 2009 .

[3]  T. Foulsham,et al.  Saliency and scan patterns in the inspection of real-world scenes: Eye movements during encoding and recognition , 2009 .

[4]  Alex D. Hwang,et al.  A model of top-down attentional control during visual search in complex scenes. , 2009, Journal of vision.

[5]  T. Foulsham,et al.  What can saliency models predict about eye movements? Spatial and sequential aspects of fixations during encoding and recognition. , 2008, Journal of vision.

[6]  Benjamin W Tatler,et al.  The central fixation bias in scene viewing: selecting an optimal viewing position independently of motor biases and image feature distributions. , 2007, Journal of vision.

[7]  Linus Holm,et al.  Memory for scenes: Refixations reflect retrieval , 2007, Memory & cognition.

[8]  Preeti Verghese,et al.  Where to look next? Eye movements reduce local uncertainty. , 2007, Journal of vision.

[9]  Antonio Torralba,et al.  Contextual guidance of eye movements and attention in real-world scenes: the role of global features in object search. , 2006, Psychological review.

[10]  Miguel P Eckstein,et al.  Attentional Cues in Real Scenes, Saccadic Targeting, and Bayesian Priors , 2005, Psychological science.

[11]  Aude Oliva,et al.  Human Learning of Contextual Priors for Object Search: Where does the time go? , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05) - Workshops.

[12]  Iain D. Gilchrist,et al.  Visual correlates of fixation selection: effects of scale and time , 2005, Vision Research.

[13]  J. Henderson Human gaze control during real-world scene perception , 2003, Trends in Cognitive Sciences.

[14]  Derrick J. Parkhurst,et al.  Scene content selected by active vision. , 2003, Spatial vision.

[15]  Eileen Kowler,et al.  Eye movements during visual search: the costs of choosing the optimal path , 2001, Vision Research.

[16]  G. Knoblich,et al.  Predicting the Effects of Actions: Interactions of Perception and Action , 2001, Psychological science.

[17]  M. Chun,et al.  Contextual Cueing: Implicit Learning and Memory of Visual Context Guides Spatial Attention , 1998, Cognitive Psychology.

[18]  D. S. Wooding,et al.  Fixation Patterns Made during Brief Examination of Two-Dimensional Images , 1997, Perception.

[19]  L. Stark,et al.  Spontaneous Eye Movements During Visual Imagery Reflect the Content of the Visual Scene , 1997, Journal of Cognitive Neuroscience.

[20]  D G Pelli,et al.  The VideoToolbox software for visual psychophysics: transforming numbers into movies. , 1997, Spatial vision.

[21]  D H Brainard,et al.  The Psychophysics Toolbox. , 1997, Spatial vision.

[22]  J. Wolfe,et al.  Guided Search 2.0 A revised model of visual search , 1994, Psychonomic bulletin & review.

[23]  E. Reed The Ecological Approach to Visual Perception , 1989 .

[24]  F. Craik,et al.  Levels of Processing in Human Memory , 1979 .

[25]  Endel Tulving,et al.  Encoding specificity and retrieval processes in episodic memory. , 1973 .

[26]  L. Stark,et al.  Scanpaths in Eye Movements during Pattern Perception , 1971, Science.

[27]  A. L. I︠A︡rbus Eye Movements and Vision , 1967 .

[28]  A. L. Yarbus,et al.  Eye Movements and Vision , 1967, Springer US.