Real-time Enhancement of Image and Video Saliency using Semantic Depth of Field

In this paper, we propose a method for automatically directing viewers’ visual attention to important regions of images and videos in low-level vision. Inspired by the modern model of visual attention, the importance map of an input scene is automatically calculated by the combination of low-level features such as intensity and color, which are extracted using spatial filters in different spatial frequencies, together with a set of temporal features extracted using a temporal filter in case of dynamic scenes. A variable-kernel-convolution based on the importance map is then performed on the input scene, in order to make semantic depth of field effects in a way that important regions remain focused while others are blurred. The pipeline of our method is efficient enough to be executed in real time on modern low-end machines, and the associated experiment demonstrates that the proposed system can be complementary to the human visual system.

[1]  Tim K Marks,et al.  SUN: A Bayesian framework for saliency using natural statistics. , 2008, Journal of vision.

[2]  Daniel J. Simons,et al.  Current Approaches to Change Blindness , 2000 .

[3]  Nuno Vasconcelos,et al.  Bottom-up saliency is a discriminant process , 2007, 2007 IEEE 11th International Conference on Computer Vision.

[4]  Silvia Miksch,et al.  Useful Properties of Semantic Depth of Field for Better F+C Visualization , 2002, VisSym.

[5]  Pietro Perona,et al.  Graph-Based Visual Saliency , 2006, NIPS.

[6]  S Ullman,et al.  Shifts in selective visual attention: towards the underlying neural circuitry. , 1985, Human neurobiology.

[7]  Pierre Baldi,et al.  Bayesian surprise attracts human attention , 2005, Vision Research.

[8]  Silvia Miksch,et al.  Semantic depth of field , 2001, IEEE Symposium on Information Visualization, 2001. INFOVIS 2001..

[9]  Liqing Zhang,et al.  Saliency Detection: A Spectral Residual Approach , 2007, 2007 IEEE Conference on Computer Vision and Pattern Recognition.

[10]  Matthew H Tong,et al.  of the Annual Meeting of the Cognitive Science Society Title SUNDAy : Saliency Using Natural Statistics for Dynamic Analysis of Scenes Permalink , 2009 .

[11]  HongJiang Zhang,et al.  Contrast-based image attention analysis by using fuzzy growing , 2003, MULTIMEDIA '03.

[12]  John K. Tsotsos Analyzing vision at the complexity level , 1990, Behavioral and Brain Sciences.

[13]  C. Koch,et al.  Computational modelling of visual attention , 2001, Nature Reviews Neuroscience.

[14]  S. Süsstrunk,et al.  Frequency-tuned salient region detection , 2009, CVPR 2009.

[15]  John K. Tsotsos,et al.  Attention based on information maximization , 2010 .

[16]  Silvia Miksch,et al.  Focus and Context Taken Literally , 2002, IEEE Computer Graphics and Applications.

[17]  Christof Koch,et al.  A Model of Saliency-Based Visual Attention for Rapid Scene Analysis , 2009 .