User-Specific Visual Attention Estimation Based on Visual Similarity and Spatial Information in Images

This paper presents a method for user-specific visual attention estimation based on visual similarities and spatial information in images. In order to estimate the user- specific visual attention, the proposed method calculates two kinds of saliency maps. One is constructed as a visual similarity-based saliency map, and the other is constructed by considering spatial information of objects in images. The proposed method performs a fusion of these two maps for considering visual similarities and spatial information. This is the biggest contribution of this paper. Therefore, improvement of the estimation performance of the user-specific visual attention is realized.

[1]  Miki Haseyama,et al.  User-centric Visual Attention Estimation Based on Relationship between Image and Eye Gaze Data , 2018, 2018 IEEE 7th Global Conference on Consumer Electronics (GCCE).

[2]  Pietro Perona,et al.  Graph-Based Visual Saliency , 2006, NIPS.

[3]  Kilian Q. Weinberger,et al.  Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[4]  Christof Koch,et al.  Image Signature: Highlighting Sparse Salient Regions , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[5]  Pietro Perona,et al.  Microsoft COCO: Common Objects in Context , 2014, ECCV.

[6]  Qi Zhao,et al.  SALICON: Saliency in Context , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[7]  Christof Koch,et al.  A Model of Saliency-Based Visual Attention for Rapid Scene Analysis , 2009 .

[8]  Shabab Bazrafkan,et al.  Eye Gaze for Consumer Electronics: Controlling and commanding intelligent systems. , 2015, IEEE Consumer Electronics Magazine.

[9]  Noel E. O'Connor,et al.  SalGAN: Visual Saliency Prediction with Generative Adversarial Networks , 2017, ArXiv.