Meaning and Attentional Guidance in Scenes: A Review of the Meaning Map Approach

Perception of a complex visual scene requires that important regions be prioritized and attentionally selected for processing. What is the basis for this selection? Although much research has focused on image salience as an important factor guiding attention, relatively little work has focused on semantic salience. To address this imbalance, we have recently developed a new method for measuring, representing, and evaluating the role of meaning in scenes. In this method, the spatial distribution of semantic features in a scene is represented as a meaning map. Meaning maps are generated from crowd-sourced responses given by naïve subjects who rate the meaningfulness of a large number of scene patches drawn from each scene. Meaning maps are coded in the same format as traditional image saliency maps, and therefore both types of maps can be directly evaluated against each other and against maps of the spatial distribution of attention derived from viewers’ eye fixations. In this review we describe our work focusing on comparing the influences of meaning and image salience on attentional guidance in real-world scenes across a variety of viewing tasks that we have investigated, including memorization, aesthetic judgment, scene description, and saliency search and judgment. Overall, we have found that both meaning and salience predict the spatial distribution of attention in a scene, but that when the correlation between meaning and salience is statistically controlled, only meaning uniquely accounts for variance in attention.

[1]  M. Tinker How People Look at Pictures. , 1936 .

[2]  Antonio Torralba,et al.  LabelMe: A Database and Web-Based Tool for Image Annotation , 2008, International Journal of Computer Vision.

[3]  S Ullman,et al.  Shifts in selective visual attention: towards the underlying neural circuitry. , 1985, Human neurobiology.

[4]  J. Henderson Human gaze control during real-world scene perception , 2003, Trends in Cognitive Sciences.

[5]  J. Henderson,et al.  High-level scene perception. , 1999, Annual review of psychology.

[6]  J. Henderson Gaze Control as Prediction , 2017, Trends in Cognitive Sciences.

[7]  D. Ballard,et al.  Eye movements in natural behavior , 2005, Trends in Cognitive Sciences.

[8]  George L. Malcolm,et al.  Searching in the dark: Cognitive relevance drives attention in real-world scenes , 2009, Psychonomic bulletin & review.

[9]  J. Wolfe,et al.  Five factors that guide attention in visual search , 2017, Nature Human Behaviour.

[10]  C. Koch,et al.  Computational modelling of visual attention , 2001, Nature Reviews Neuroscience.

[11]  K. Rayner The 35th Sir Frederick Bartlett Lecture: Eye movements and attention in reading, scene perception, and visual search , 2009, Quarterly journal of experimental psychology.

[12]  M. Pomplun,et al.  Guidance of visual attention by semantic information in real-world scenes , 2014, Front. Psychol..

[13]  J. Antes The time course of picture viewing. , 1974, Journal of experimental psychology.

[14]  Daniel F. Parks,et al.  Complementary effects of gaze direction and early saliency in guiding fixations during free viewing. , 2014, Journal of vision.

[15]  Christof Koch,et al.  A Model of Saliency-Based Visual Attention for Rapid Scene Analysis , 2009 .

[16]  Taylor R. Hayes,et al.  Meaning guides attention during scene viewing, even when it is irrelevant , 2018, Attention, perception & psychophysics.

[17]  K. Turano,et al.  Oculomotor strategies for the direction of gaze tested with a real-world activity , 2003, Vision Research.

[18]  M. Hayhoe,et al.  In what ways do eye movements contribute to everyday activities? , 2001, Vision Research.

[19]  Leon A. Gatys,et al.  Understanding Low- and High-Level Contributions to Fixation Prediction , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[20]  N. Mackworth,et al.  The gaze selects informative details within pictures , 1967 .

[21]  Taylor R. Hayes,et al.  Meaning guides attention in real-world scene images: Evidence from eye movements and meaning maps , 2017, bioRxiv.

[22]  T. Foulsham,et al.  How Does the Purpose of Inspection Influence the Potency of Visual Salience in Scene Perception? , 2007, Perception.

[23]  F. Ferreira,et al.  How incremental is language production? Evidence from the production of utterances requiring the computation of arithmetic sums , 2002 .

[24]  D. Ballard,et al.  Modeling Task Control of Eye Movements , 2014, Current Biology.

[25]  P. de Graef,et al.  Perceptual effects of scene context on object identification , 1990, Psychological research.

[26]  Gwendolyn Rehrig,et al.  Meaning Guides Attention during Real-World Scene Description , 2018, Scientific Reports.

[27]  A. Treisman,et al.  A feature-integration theory of attention , 1980, Cognitive Psychology.

[28]  D. Ballard,et al.  Eye guidance in natural vision: reinterpreting salience. , 2011, Journal of vision.

[29]  J. Henderson,et al.  Prioritizing new objects for eye fixation in real-world scenes: Effects of object–scene consistency , 2008 .

[30]  Eyal M. Reingold,et al.  Direct control of fixation times in scene viewing: Evidence from analysis of the distribution of first fixation duration , 2012 .

[31]  J. Henderson Regarding Scenes , 2007 .

[32]  Mary M Hayhoe,et al.  Task and context determine where you look. , 2016, Journal of vision.

[33]  N. Mackworth,et al.  Cognitive determinants of fixation location during picture viewing. , 1978, Journal of experimental psychology. Human perception and performance.

[34]  J. Wolfe,et al.  Guided Search 2.0 A revised model of visual search , 1994, Psychonomic bulletin & review.

[35]  C. Koch,et al.  Task-demands can immediately reverse the effects of sensory-driven saliency in complex visual stimuli. , 2008, Journal of vision.

[36]  Ali Borji,et al.  State-of-the-Art in Visual Attention Modeling , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[37]  Antje Nuthmann,et al.  Eye movement control during scene viewing: immediate effects of scene luminance on fixation durations. , 2013, Journal of experimental psychology. Human perception and performance.

[38]  Antonio Torralba,et al.  Contextual guidance of eye movements and attention in real-world scenes: the role of global features in object search. , 2006, Psychological review.

[39]  A. L. Yarbus,et al.  Eye Movements and Vision , 1967, Springer US.

[40]  J. Henderson,et al.  CRISP: a computational model of fixation durations in scene viewing. , 2010, Psychological review.

[41]  Ali Borji,et al.  Quantitative Analysis of Human-Model Agreement in Visual Saliency Modeling: A Comparative Study , 2013, IEEE Transactions on Image Processing.

[42]  S. Liversedge,et al.  Saccadic eye movements and cognition , 2000, Trends in Cognitive Sciences.

[43]  J. Henderson,et al.  The effects of semantic consistency on eye movements during complex scene viewing , 1999 .

[44]  Falk Huettig,et al.  The tug of war between phonological, semantic and shape information in language-mediated visual search , 2007 .

[45]  Michael L. Mack,et al.  Viewing task influences eye movement control during active scene perception. , 2009, Journal of vision.

[46]  Graham L. Pierce,et al.  Eye movements during scene viewing: Evidence for mixed control of fixation durations , 2008, Psychonomic bulletin & review.

[47]  K. Rayner,et al.  Eye movements and scene perception. , 1992, Canadian journal of psychology.

[48]  Michael L. Mack,et al.  VISUAL SALIENCY DOES NOT ACCOUNT FOR EYE MOVEMENTS DURING VISUAL SEARCH IN REAL-WORLD SCENES , 2007 .

[49]  Taylor R. Hayes,et al.  Meaning-based guidance of attention in scenes as revealed by meaning maps , 2017, Nature Human Behaviour.

[50]  George L. Malcolm,et al.  Intrusive effects of task-irrelevant information on visual selective attention: semantics and size. , 2019, Current opinion in psychology.

[51]  P M van Diepen,et al.  Brief foveal masking during scene perception. , 1999, Acta psychologica.

[52]  Steven G. Luke,et al.  Eye movement control in scene viewing and reading: evidence from the stimulus onset delay paradigm. , 2013, Journal of experimental psychology. Human perception and performance.

[53]  Pietro Perona,et al.  Graph-Based Visual Saliency , 2006, NIPS.

[54]  J. Henderson,et al.  Does gravity matter? Effects of semantic and syntactic inconsistencies on the allocation of attention during scene perception. , 2009, Journal of vision.

[55]  Gregory J. Zelinsky,et al.  Scene context guides eye movements during visual search , 2006, Vision Research.

[56]  J. Henderson,et al.  How are eye fixation durations controlled during scene viewing? Further evidence from a scene onset delay paradigm , 2009 .

[57]  Derrick J. Parkhurst,et al.  Modeling the role of salience in the allocation of overt visual attention , 2002, Vision Research.