A new approach to modeling the influence of image features on fixation selection in scenes

Which image characteristics predict where people fixate when memorizing natural images? To answer this question, we introduce a new analysis approach that combines a novel scene‐patch analysis with generalized linear mixed models (GLMMs). Our method allows for (1) directly describing the relationship between continuous feature value and fixation probability, and (2) assessing each feature's unique contribution to fixation selection. To demonstrate this method, we estimated the relative contribution of various image features to fixation selection: luminance and luminance contrast (low‐level features); edge density (a mid‐level feature); visual clutter and image segmentation to approximate local object density in the scene (higher‐level features). An additional predictor captured the central bias of fixation. The GLMM results revealed that edge density, clutter, and the number of homogenous segments in a patch can independently predict whether image patches are fixated or not. Importantly, neither luminance nor contrast had an independent effect above and beyond what could be accounted for by the other predictors. Since the parcellation of the scene and the selection of features can be tailored to the specific research question, our approach allows for assessing the interplay of various factors relevant for fixation selection in scenes in a powerful and flexible manner.

[1]  A. Mizuno,et al.  A change of the leading player in flow Visualization technique , 2006, J. Vis..

[2]  Dorin Comaniciu,et al.  Mean Shift: A Robust Approach Toward Feature Space Analysis , 2002, IEEE Trans. Pattern Anal. Mach. Intell..

[3]  J. Henderson,et al.  Object-based attentional selection in scene viewing. , 2010, Journal of vision.

[4]  Asha Iyer,et al.  Components of bottom-up gaze allocation in natural images , 2005, Vision Research.

[5]  Leslie S. Smith,et al.  The principal components of natural images , 1992 .

[6]  Robert A. Frazor,et al.  Independence of luminance and contrast in natural scenes and in the early visual system , 2005, Nature Neuroscience.

[7]  H. Deubel,et al.  Saccade target selection and object recognition: Evidence for a common attentional mechanism , 1996, Vision Research.

[8]  Bernhard Schölkopf,et al.  Implicit Wiener Series for Higher-Order Image Analysis , 2004, NIPS.

[9]  Derrick J. Parkhurst,et al.  Texture contrast attracts overt visual attention in natural scenes , 2004, The European journal of neuroscience.

[10]  Christof Koch,et al.  Modeling attention to salient proto-objects , 2006, Neural Networks.

[11]  Derrick J. Parkhurst,et al.  Modeling the role of salience in the allocation of overt visual attention , 2002, Vision Research.

[12]  B. Moulden,et al.  The Standard Deviation of Luminance as a Metric for Contrast in Random-Dot Images , 1990, Perception.

[13]  Jacob Cohen,et al.  Applied multiple regression/correlation analysis for the behavioral sciences , 1979 .

[14]  G. Hauske,et al.  Object and scene analysis by saccadic eye-movements: an investigation with higher-order statistics. , 2000, Spatial vision.

[15]  Katsumi Aoki,et al.  Recent development of flow visualization , 2004, J. Vis..

[16]  Peter Meer,et al.  Synergism in low level vision , 2002, Object recognition supported by user interaction for service robots.

[17]  Maciej Pajak,et al.  Object-based saccadic selection during scene perception: evidence from viewing position effects. , 2013, Journal of vision.

[18]  Peter Meer,et al.  Edge Detection with Embedded Confidence , 2001, IEEE Trans. Pattern Anal. Mach. Intell..

[19]  Alessandra S. Souza,et al.  Refreshing memory traces: thinking of an item improves retrieval from visual working memory , 2015, Annals of the New York Academy of Sciences.

[20]  T. Smith,et al.  Attentional synchrony and the influence of viewing task on gaze behavior in static and dynamic scenes. , 2013, Journal of vision.

[21]  B. Tatler,et al.  Deriving an appropriate baseline for describing fixation behaviour , 2014, Vision Research.

[22]  Kerry Hourigan,et al.  Wake transition of a rolling sphere , 2011, J. Vis..

[23]  T. Jaeger,et al.  Categorical Data Analysis: Away from ANOVAs (transformation or not) and towards Logit Mixed Models. , 2008, Journal of memory and language.

[24]  Scott T Grafton,et al.  Multimodal neuroimaging evidence linking memory and attention systems during visual search cued by context , 2015, Annals of the New York Academy of Sciences.

[25]  Ralf Engbert,et al.  Length, frequency, and predictability effects of words on eye movements in reading , 2004 .

[26]  Derrick J. Parkhurst,et al.  Scene content selected by active vision. , 2003, Spatial vision.

[27]  Benjamin W Tatler,et al.  The central fixation bias in scene viewing: selecting an optimal viewing position independently of motor biases and image feature distributions. , 2007, Journal of vision.

[28]  Iain D. Gilchrist,et al.  Visual correlates of fixation selection: effects of scale and time , 2005, Vision Research.

[29]  D. Barr,et al.  Random effects structure for confirmatory hypothesis testing: Keep it maximal. , 2013, Journal of memory and language.

[30]  P. Perona,et al.  Objects predict fixations better than early saliency. , 2008, Journal of vision.

[31]  D. Ballard,et al.  Eye guidance in natural vision: reinterpreting salience. , 2011, Journal of vision.

[32]  Dimitris Samaras,et al.  Modeling visual clutter perception using proto-object segmentation. , 2014, Journal of vision.

[33]  D. S. Wooding,et al.  Fixation Patterns Made during Brief Examination of Two-Dimensional Images , 1997, Perception.

[34]  Ronald A. Rensink The Dynamic Representation of Scenes , 2000 .

[35]  Myriam Chanceaux,et al.  The influence of clutter on real-world scene search: evidence from search efficiency and eye movements. , 2009, Journal of vision.

[36]  Peter König,et al.  Measures and Limits of Models of Fixation Selection , 2011, PloS one.

[37]  Nao Ninomiya,et al.  The 10th anniversary of journal of visualization , 2007, J. Vis..

[38]  Stefan Mihalas,et al.  A model of proto-object based saliency , 2014, Vision Research.

[39]  R. Baddeley,et al.  Do we look at lights? Using mixture modelling to distinguish between low- and high-level factors in natural image viewing , 2009 .

[40]  Frédo Durand,et al.  Learning to predict where humans look , 2009, 2009 IEEE 12th International Conference on Computer Vision.

[41]  P. König,et al.  Does luminance‐contrast contribute to a saliency map for overt visual attention? , 2003, The European journal of neuroscience.

[42]  D. Bates,et al.  Linear Mixed-Effects Models using 'Eigen' and S4 , 2015 .

[43]  Yuanzhen Li,et al.  Measuring visual clutter. , 2007, Journal of vision.

[44]  R. Harald Baayen,et al.  Analyzing linguistic data: a practical introduction to statistics using R, 1st Edition , 2008 .

[45]  C. Koch,et al.  Computational modelling of visual attention , 2001, Nature Reviews Neuroscience.

[46]  D. Barr Analyzing ‘visual world’ eyetracking data using multilevel logistic regression , 2008 .

[47]  A. Agresti,et al.  Approximate is Better than “Exact” for Interval Estimation of Binomial Proportions , 1998 .

[48]  Ali Borji,et al.  State-of-the-Art in Visual Attention Modeling , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[49]  Alexander C. Schütz,et al.  Eye movements and perception: a selective review. , 2011, Journal of vision.

[50]  W. Einhäuser,et al.  Overt attention in natural scenes: Objects dominate features , 2015, Vision Research.

[51]  Reinhold Kliegl,et al.  Semantic preview benefit during reading. , 2014, Journal of experimental psychology. Learning, memory, and cognition.

[52]  Roland J. Baddeley,et al.  High frequency edges (but not contrast) predict where we fixate: A Bayesian system identification analysis , 2006, Vision Research.

[53]  D J Field,et al.  Relations between the statistics of natural images and the response properties of cortical cells. , 1987, Journal of the Optical Society of America. A, Optics and image science.

[54]  P Reinagel,et al.  Natural scene statistics at the centre of gaze. , 1999, Network.

[55]  Masaaki Kawahashi,et al.  Renovation of Journal of Visualization , 2010, J. Vis..

[56]  D. S. Wooding,et al.  The relationship between the locations of spatial features and those of fixations made during visual examination of briefly presented images. , 1996, Spatial vision.

[57]  Austin F. Frank,et al.  Analyzing linguistic data: a practical introduction to statistics using R , 2010 .

[58]  K. Fujii,et al.  Visualization for the analysis of fluid motion , 2005, J. Vis..