An expected motion information model of salience for active cameras

Faced with the challenge of learning about its environment, an important first step for a robot is deciding how to direct its sensors to extract meaningful information. Computational models of visual salience have been developed to predict where humans tend to look in visual scenes, and thus they may provide useful heuristics for orienting robotic cameras. However, as we show here, current models of visual salience exhibit some important problems when applied to active robotic cameras. Here, we describe a new model of visual salience, named EMI, specifically designed to work on robotic cameras. The intuition behind this model is that, regardless of the task at hand, it is critical for robot cameras to keep track of motion, be it caused by camera movement or by world movement. Thus, it is reasonable for cameras to focus on image regions that are expected to provide high information about future motion, and to do so in a way that blurs the image as little as possible. We show that EMI predicts human fixations at least as well as current models of visual salience. In addition, we show that EMI overcomes the limitations that current visual salience models have when applied to robotic, active cameras.

[1]  Javier R. Movellan,et al.  Tracking Motion, Deformation, and Texture Using Conditionally Gaussian Processes , 2010, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[2]  Harish Katti,et al.  An Eye Fixation Database for Saliency Detection in Images , 2010, ECCV.

[3]  Tim K Marks,et al.  SUN: A Bayesian framework for saliency using natural statistics. , 2008, Journal of vision.

[4]  Christof Koch,et al.  A Model of Saliency-Based Visual Attention for Rapid Scene Analysis , 2009 .

[5]  Jitendra Malik,et al.  An Information Maximization Model of Eye Movements , 2004, NIPS.

[6]  John K. Tsotsos,et al.  Saliency Based on Information Maximization , 2005, NIPS.

[7]  Antonio Torralba,et al.  Contextual Modulation of Target Saliency , 2001, NIPS.

[8]  Wilson S. Geisler,et al.  Optimal eye movement strategies in visual search , 2005, Nature.

[9]  M. Asada,et al.  Visual attention by saliency leads cross-modal body representation , 2008, 2008 7th IEEE International Conference on Development and Learning.

[10]  Giulio Sandini,et al.  Learning VOR-like stabilization reflexes in robots , 2000, ESANN.

[11]  D. Ballard,et al.  Eye guidance in natural vision: reinterpreting salience. , 2011, Journal of vision.

[12]  Nuno Vasconcelos,et al.  The discriminant center-surround hypothesis for bottom-up saliency , 2007, NIPS.

[13]  C. Koch,et al.  Computational modelling of visual attention , 2001, Nature Reviews Neuroscience.

[14]  Garrison W. Cottrell,et al.  Visual saliency model for robot cameras , 2008, 2008 IEEE International Conference on Robotics and Automation.

[15]  N.J. Butko,et al.  I-POMDP: An infomax model of eye movement , 2008, 2008 7th IEEE International Conference on Development and Learning.

[16]  Bernhard Schölkopf,et al.  A Nonparametric Approach to Bottom-Up Visual Saliency , 2006, NIPS.