Applying computational tools to predict gaze direction in interactive visual environments

Future interactive virtual environments will be “attention-aware,” capable of predicting, reacting to, and ultimately influencing the visual attention of their human operators. Before such environments can be realized, it is necessary to operationalize our understanding of the relevant aspects of visual perception, in the form of fully automated computational heuristics that can efficiently identify locations that would attract human gaze in complex dynamic environments. One promising approach to designing such heuristics draws on ideas from computational neuroscience. We compared several neurobiologically inspired heuristics with eye-movement recordings from five observers playing video games, and found that human gaze was better predicted by heuristics that detect outliers from the global distribution of visual features than by purely local heuristics. Heuristics sensitive to dynamic events performed best overall. Further, heuristic prediction power differed more between games than between different human observers. While other factors clearly also influence eye position, our findings suggest that simple neurally inspired algorithmic methods can account for a significant portion of human gaze behavior in a naturalistic, interactive setting. These algorithms may be useful in the implementation of interactive virtual environments, both to predict the cognitive state of human operators, as well as to effectively endow virtual agents in the system with humanlike visual behavior.

[1]  M. Hayhoe Advances in Relating Eye Movements and Cognition. , 2004, Infancy : the official journal of the International Society on Infant Studies.

[2]  Derrick J. Parkhurst,et al.  Texture contrast attracts overt visual attention in natural scenes , 2004, The European journal of neuroscience.

[3]  E. Peli,et al.  Scanpaths of motion sequences : where people look when watching movies , 2005 .

[4]  Dave M. Stampe,et al.  Heuristic filtering and reliable calibration methods for video-based pupil-tracking systems , 1993 .

[5]  M. Hayhoe,et al.  In what ways do eye movements contribute to everyday activities? , 2001, Vision Research.

[6]  A. Treisman,et al.  A feature-integration theory of attention , 1980, Cognitive Psychology.

[7]  S Ullman,et al.  Shifts in selective visual attention: towards the underlying neural circuitry. , 1985, Human neurobiology.

[8]  Pierre Baldi,et al.  A principled approach to detecting surprising events in video , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[9]  R. E. Parker Picture processing during recognition. , 1978, Journal of experimental psychology. Human perception and performance.

[10]  A. L. Yarbus Eye Movements During Perception of Complex Objects , 1967 .

[11]  M. Pomplun Saccadic selectivity in complex visual search displays , 2006, Vision Research.

[12]  Christopher E. Peters,et al.  Bottom-up visual attention for virtual human animation , 2003, Proceedings 11th IEEE International Workshop on Program Comprehension.

[13]  Christof Koch,et al.  A Model of Saliency-Based Visual Attention for Rapid Scene Analysis , 2009 .

[14]  L. Itti,et al.  Modeling the influence of task on attention , 2005, Vision Research.

[15]  Jochen Triesch,et al.  Vision in natural and virtual environments , 2002, ETRA.

[16]  S A Finney,et al.  Real-time data collection in Linux: A case study , 2001, Behavior research methods, instruments, & computers : a journal of the Psychonomic Society, Inc.

[17]  Bryan Reimer,et al.  On-road driver eye movement tracking using head-mounted devices , 2002, ETRA.

[18]  Claudio M. Privitera,et al.  Algorithms for Defining Visual Regions-of-Interest: Comparison with Eye Fixations , 2000, IEEE Trans. Pattern Anal. Mach. Intell..

[19]  V. Tosi,et al.  Scanning eye movements made when viewing film: preliminary observations. , 1997, The International journal of neuroscience.

[20]  Philip J. Barnard,et al.  Using Film Cutting Techniques in Interface Design , 2003, Hum. Comput. Interact..

[21]  D. Ballard,et al.  Memory Representations in Natural Tasks , 1995, Journal of Cognitive Neuroscience.

[22]  Wilson S. Geisler,et al.  Optimal eye movement strategies in visual search , 2005, Nature.

[23]  K. Rayner Eye movements in reading and information processing: 20 years of research. , 1998, Psychological bulletin.

[24]  Laurent Itti,et al.  Bottom-up and top-down influences on attentional allocation in natural dynamic scenes , 2004 .

[25]  Edward H. Adelson,et al.  The Laplacian Pyramid as a Compact Image Code , 1983, IEEE Trans. Commun..

[26]  S. Vogt Looking at Paintings: Patterns of Eye Movements in Artistically Nave and Sophisticated Subjects , 1999, Leonardo.

[27]  Antonio Torralba,et al.  Modeling global scene factors in attention. , 2003, Journal of the Optical Society of America. A, Optics, image science, and vision.

[28]  Laurent Itti,et al.  Realistic avatar eye and head animation using a neurobiological model of visual attention , 2004, SPIE Optics + Photonics.

[29]  Rajesh P. N. Rao,et al.  Eye movements in iconic visual search , 2002, Vision Research.

[30]  Christof Koch,et al.  Feature combination strategies for saliency-based visual attention systems , 2001, J. Electronic Imaging.

[31]  Demetri Terzopoulos,et al.  Animat vision: Active vision in artificial animals , 1995, Proceedings of IEEE International Conference on Computer Vision.

[32]  P Reinagel,et al.  Natural scene statistics at the centre of gaze. , 1999, Network.

[33]  M. Land,et al.  The Roles of Vision and Eye Movements in the Control of Activities of Daily Living , 1998, Perception.

[34]  J. Henderson,et al.  High-level scene perception. , 1999, Annual review of psychology.

[35]  Claudio M. Privitera,et al.  Evaluating image processing algorithms that predict regions of interest , 1998, Pattern Recognit. Lett..

[36]  Gerhard Krieger,et al.  Investigation of a sensorimotor system for saccadic scene analysis: an integrated approach , 1998 .

[37]  Asha Iyer,et al.  Components of bottom-up gaze allocation in natural images , 2005, Vision Research.

[38]  Derrick J. Parkhurst,et al.  Modeling the role of salience in the allocation of overt visual attention , 2002, Vision Research.

[39]  Mary M Hayhoe,et al.  Visual memory and motor planning in a natural task. , 2003, Journal of vision.

[40]  J. Bailenson,et al.  Digital Chameleons , 2005, Psychological science.

[41]  D. S. Wooding,et al.  Fixation Patterns Made during Brief Examination of Two-Dimensional Images , 1997, Perception.

[42]  Ronald A. Rensink The Dynamic Representation of Scenes , 2000 .

[43]  J. Wolfe,et al.  What attributes guide the deployment of visual attention and how do they do it? , 2004, Nature Reviews Neuroscience.

[44]  Laurent Itti,et al.  Optimal cue selection strategy , 2005, NIPS.

[45]  Alexander Toet,et al.  Gaze directed displays as an enabling technology for attention aware systems , 2006, Comput. Hum. Behav..

[46]  Pietro Perona,et al.  Overcomplete steerable pyramid filters and rotation invariance , 1994, 1994 Proceedings of IEEE Conference on Computer Vision and Pattern Recognition.