Object co-occurrence serves as a contextual cue to guide and facilitate visual search in a natural viewing environment.

There is accumulating evidence that scene context can guide and facilitate visual search (e.g., A. Torralba, A. Oliva, M. S. Castelhano, & J. M. Henderson, 2006). Previous studies utilized stimuli of restricted size, a fixed head position, and context defined by the global spatial configuration of the scene. Thus, it is unknown whether similar effects generalize to natural viewing environments and to context defined by local object co-occurrence. Here, with a mobile eye tracker, we investigated the effects of object co-occurrence on search performance under naturalistic conditions. Observers searched for low-visibility target objects on tables cluttered with everyday objects. Targets were either located adjacent to larger, more visible "cue" objects that they regularly co-occurred in natural scenes (expected condition) or elsewhere in the display, surrounded by unrelated objects (unexpected condition). Mean search times were shorter for targets at expected locations as compared to unexpected locations. Additionally, context guided eye movements, as more fixations were directed toward cue objects than other non-target objects, particularly when the cue was contextually relevant to the current search target. These results could not be accounted for by image saliency models. Thus, we conclude that object co-occurrence can serve as a contextual cue to facilitate search and guide eye movements in natural environments.

[1]  C. Koch,et al.  A saliency-based search mechanism for overt and covert shifts of visual attention , 2000, Vision Research.

[2]  M. Chun,et al.  Implicit, long-term spatial contextual memory. , 2003, Journal of experimental psychology. Learning, memory, and cognition.

[3]  R. D. Gordon,et al.  Attention to smoking-related and incongruous objects during scene viewing. , 2008, Acta psychologica.

[4]  M. Land Eye Movements in Daily Life , 2003 .

[5]  R. Rosenholtz A simple saliency model predicts a number of motion popout phenomena , 1999, Vision Research.

[6]  J. P. Thomas,et al.  A signal detection model predicts the effects of set size on visual search accuracy for feature, conjunction, triple conjunction, and disjunction displays , 2000, Perception & psychophysics.

[7]  M. Bar Visual objects in context , 2004, Nature Reviews Neuroscience.

[8]  Mark W. Becker,et al.  Object-intrinsic oddities draw early saccades. , 2007, Journal of experimental psychology. Human perception and performance.

[9]  M. Chun,et al.  Top-Down Attentional Guidance Based on Implicit Learning of Visual Covariation , 1999 .

[10]  Miguel P Eckstein,et al.  Attentional Cues in Real Scenes, Saccadic Targeting, and Bayesian Priors , 2005, Psychological science.

[11]  Michael L. Mack,et al.  Human Gaze Control in RealWorld Search , 2004, WAPCV.

[12]  James R. Brockmole,et al.  Using real-world scenes as contextual cues for search , 2006 .

[13]  A. Kramer,et al.  Attentional guidance of the eyes by contextual information and abrupt onsets. , 2001, Perception & psychophysics.

[14]  Rajesh P. N. Rao,et al.  Eye movements in iconic visual search , 2002, Vision Research.

[15]  John M Henderson,et al.  The Role of Meaning in Contextual Cueing: Evidence from Chess Expertise , 2008, Quarterly journal of experimental psychology.

[16]  John K. Tsotsos,et al.  Saliency, attention, and visual search: an information theoretic approach. , 2009, Journal of vision.

[17]  G. Zelinsky A theory of eye movements during target acquisition. , 2008, Psychological review.

[18]  D. Ballard,et al.  Eye movements in natural behavior , 2005, Trends in Cognitive Sciences.

[19]  Jason A. Droll,et al.  Expected object position of two hundred fifty observers predicts first fixations of seventy seven separate observers during search , 2010 .

[20]  M. Chun,et al.  Contextual Cueing: Implicit Learning and Memory of Visual Context Guides Spatial Attention , 1998, Cognitive Psychology.

[21]  Miguel P. Eckstein,et al.  Gaze control and memory for objects while walking in a real world environment , 2009 .

[22]  J. Henderson,et al.  The effects of semantic consistency on eye movements during complex scene viewing , 1999 .

[23]  Christof Koch,et al.  Selecting One Among the Many: A Simple Network Implementing Shifts in Selective Visual Attention , 1984 .

[24]  George L. Malcolm,et al.  Searching in the dark: Cognitive relevance drives attention in real-world scenes , 2009, Psychonomic bulletin & review.

[25]  J. Henderson,et al.  Initial scene representations facilitate eye movement guidance in visual search. , 2007, Journal of experimental psychology. Human perception and performance.

[26]  Alain Berthoz,et al.  The Head-neck sensory motor system , 1992 .

[27]  Yuhong V. Jiang,et al.  What is learned in spatial contextual cuing— configuration or individual locations? , 2004, Perception & psychophysics.

[28]  Tim K Marks,et al.  SUN: A Bayesian framework for saliency using natural statistics. , 2008, Journal of vision.

[29]  Zygmunt Pizlo,et al.  Coordination of Head and Eyes during the Performance of Natural (and Unnatural) Visual Tasks , 2012 .

[30]  J. Findlay Saccade Target Selection During Visual Search , 1997, Vision Research.

[31]  M. Castelhano,et al.  The relative contribution of scene context and target features to visual search in scenes , 2010, Attention, perception & psychophysics.

[32]  Krista A. Ehinger,et al.  The role of color in visual search in real-world scenes: Evidence from contextual cuing , 2008, Perception & psychophysics.

[33]  James R. Brockmole,et al.  Contextual cueing in naturalistic scenes: Global and local contexts. , 2006, Journal of experimental psychology. Learning, memory, and cognition.

[34]  Timothy F. Brady,et al.  Spatial constraints on learning in visual search: modeling contextual cuing. , 2007, Journal of experimental psychology. Human perception and performance.

[35]  Michael L. Mack,et al.  VISUAL SALIENCY DOES NOT ACCOUNT FOR EYE MOVEMENTS DURING VISUAL SEARCH IN REAL-WORLD SCENES , 2007 .

[36]  L. Chalupa,et al.  The visual neurosciences , 2004 .

[37]  C. Koch,et al.  Computational modelling of visual attention , 2001, Nature Reviews Neuroscience.

[38]  A. Hollingworth Two forms of scene memory guide visual search: Memory for scene context and memory for the binding of target object to scene location , 2009 .

[39]  John K. Tsotsos,et al.  Proceedings of the Second international conference on Attention and Performance in Computational Vision , 2004 .

[40]  Miguel P Eckstein,et al.  Learning cue validity through performance feedback. , 2009, Journal of Vision.

[41]  Iain D. Gilchrist,et al.  Visual correlates of fixation selection: effects of scale and time , 2005, Vision Research.

[42]  Antonio Torralba,et al.  Modeling global scene factors in attention. , 2003, Journal of the Optical Society of America. A, Optics, image science, and vision.

[43]  Krista A. Ehinger,et al.  Modelling search for people in 900 scenes: A combined source model of eye guidance , 2009 .

[44]  M. Chun,et al.  Perceptual constraints on implicit learning of spatial context , 2002 .

[45]  J. Henderson Human gaze control during real-world scene perception , 2003, Trends in Cognitive Sciences.

[46]  M. Hayhoe,et al.  In what ways do eye movements contribute to everyday activities? , 2001, Vision Research.

[47]  Antonio Torralba,et al.  Contextual guidance of eye movements and attention in real-world scenes: the role of global features in object search. , 2006, Psychological review.

[48]  Jodi L. Davenport Consistency effects between objects in scenes , 2007, Memory & cognition.

[49]  N. Mackworth,et al.  Cognitive determinants of fixation location during picture viewing. , 1978, Journal of experimental psychology. Human perception and performance.

[50]  A. Bovik,et al.  Visual search in noise: revealing the influence of structural cues by gaze-contingent classification image analysis. , 2006, Journal of vision.

[51]  Benjamin W. Tatler,et al.  Current understanding of eye guidance , 2009 .

[52]  James R. Brockmole,et al.  Short Article: Recognition and Attention Guidance during Contextual Cueing in Real-World Scenes: Evidence from Eye Movements , 2006, Quarterly journal of experimental psychology.

[53]  Antonio Torralba,et al.  Top-down control of visual attention in object detection , 2003, Proceedings 2003 International Conference on Image Processing (Cat. No.03CH37429).

[54]  J. Henderson,et al.  Does gravity matter? Effects of semantic and syntactic inconsistencies on the allocation of attention during scene perception. , 2009, Journal of vision.

[55]  George L. Malcolm,et al.  Combining top-down processes to guide eye movements during real-world scene search. , 2010, Journal of vision.

[56]  Christof Koch,et al.  A Model of Saliency-Based Visual Attention for Rapid Scene Analysis , 2009 .

[57]  T. Foulsham,et al.  Quarterly Journal of Experimental Psychology: in press Visual saliency and semantic incongruency influence eye movements when , 2022 .

[58]  Gregory J. Zelinsky,et al.  Scene context guides eye movements during visual search , 2006, Vision Research.

[59]  Derrick J. Parkhurst,et al.  Modeling the role of salience in the allocation of overt visual attention , 2002, Vision Research.

[60]  Mary M Hayhoe,et al.  Visual memory and motor planning in a natural task. , 2003, Journal of vision.

[61]  Wilson S. Geisler,et al.  Optimal eye movement strategies in visual search , 2005, Nature.

[62]  J. Henderson,et al.  Prioritizing new objects for eye fixation in real-world scenes: Effects of object–scene consistency , 2008 .

[63]  Aude Oliva,et al.  Human Learning of Contextual Priors for Object Search: Where does the time go? , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05) - Workshops.

[64]  Miguel P Eckstein,et al.  Similar Neural Representations of the Target for Saccades and Perception during Search , 2007, The Journal of Neuroscience.

[65]  P. de Graef,et al.  Perceptual effects of scene context on object identification , 1990, Psychological research.

[66]  Miguel P Eckstein,et al.  Saccadic and perceptual performance in visual search tasks. I. Contrast detection and discrimination. , 2003, Journal of the Optical Society of America. A, Optics, image science, and vision.

[67]  Michael F. Land,et al.  The coordination of rotations of the eyes, head and trunk in saccadic turns produced in natural situations , 2004, Experimental Brain Research.

[68]  A. Bovik,et al.  An efficient technique for revealing visual search strategies with classification images , 2007, Perception & psychophysics.

[69]  D. Lamy,et al.  Task-irrelevant stimulus salience affects visual search , 2009, Vision Research.

[70]  M. Chun,et al.  Contextual cueing of visual attention , 2022 .