Visual attention analysis and prediction on human faces

Abstract Human faces are almost always the focus of visual attention because of the rich semantic information therein. While some visual attention models incorporating face cues indeed perform better in images with faces, yet there is no systematic analysis of the deployment of visual attention on human faces in the context of visual attention modelling, nor is there any specific attention model designed for face images. On faces, many high-level factors have influence on visual attention. To investigate visual attention on human faces, we first construct a Visual Attention database for Faces (VAF database), which is composed of 481 face images along with eye-tracking data of 22 viewers. Statistics of the eye-movement data show that some high-level factors such as face size, facial features and face pose have impact on visual attention. Thus we propose to build visual attention models specifically for face images through combining low-level saliency calculated by traditional saliency models with high-level facial features. Efficiency of the built models is verified on the VAF database. When combined with high-level facial features, most saliency models can achieve better performance.

[1]  Laurent Itti,et al.  Ieee Transactions on Pattern Analysis and Machine Intelligence 1 Rapid Biologically-inspired Scene Classification Using Features Shared with Visual Attention , 2022 .

[2]  Stan Sclaroff,et al.  Saliency Detection: A Boolean Map Approach , 2013, 2013 IEEE International Conference on Computer Vision.

[3]  Christof Koch,et al.  A Model of Saliency-Based Visual Attention for Rapid Scene Analysis , 2009 .

[4]  Xiongkuo Min,et al.  A hierarchical saliency detection approach for bokeh images , 2015, 2015 IEEE 17th International Workshop on Multimedia Signal Processing (MMSP).

[5]  Jing Zhang,et al.  Image retrieval using the extended salient region , 2017, Inf. Sci..

[6]  Xiongkuo Min,et al.  Visual attention on human face , 2015, 2015 Visual Communications and Image Processing (VCIP).

[7]  Pietro Perona,et al.  Graph-Based Visual Saliency , 2006, NIPS.

[8]  Soohyung Kim,et al.  Salient object detection using recursive regional feature clustering , 2017, Inf. Sci..

[9]  Aykut Erdem,et al.  Visual saliency estimation by nonlinearly integrating features using region covariances. , 2013, Journal of vision.

[10]  Guangtao Zhai,et al.  Visual saliency model based on minimum description length , 2016, 2016 IEEE International Symposium on Circuits and Systems (ISCAS).

[11]  Ali Borji,et al.  State-of-the-Art in Visual Attention Modeling , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[12]  Liming Zhang,et al.  Spatio-temporal Saliency detection using phase spectrum of quaternion fourier transform , 2008, 2008 IEEE Conference on Computer Vision and Pattern Recognition.

[13]  Stefanos Zafeiriou,et al.  Robust Discriminative Response Map Fitting with Constrained Local Models , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition.

[14]  Tim K Marks,et al.  SUN: A Bayesian framework for saliency using natural statistics. , 2008, Journal of vision.

[15]  A. Kingstone,et al.  Human Social Attention , 2009, Annals of the New York Academy of Sciences.

[16]  Frédo Durand,et al.  Learning to predict where humans look , 2009, 2009 IEEE 12th International Conference on Computer Vision.

[17]  Xiongkuo Min,et al.  Influence of compression artifacts on visual attention , 2014, 2014 IEEE International Conference on Multimedia and Expo (ICME).

[18]  Weisi Lin,et al.  A Fast Reliable Image Quality Predictor by Fusing Micro- and Macro-Structures , 2017, IEEE Transactions on Industrial Electronics.

[19]  Ramis Örlü,et al.  The life of a vortex in an axisymmetric jet , 2011, J. Vis..

[20]  Ke Gu,et al.  No-Reference Quality Assessment of Screen Content Pictures , 2017, IEEE Transactions on Image Processing.

[21]  Asha Iyer,et al.  Components of bottom-up gaze allocation in natural images , 2005, Vision Research.

[22]  Xiongkuo Min,et al.  Visual attention data for image quality assessment databases , 2014, 2014 IEEE International Symposium on Circuits and Systems (ISCAS).

[23]  Baihua Li,et al.  Quality assessment metric of stereo images considering cyclopean integration and visual saliency , 2016, Inf. Sci..

[24]  A. Kingstone,et al.  Saliency does not account for fixations to eyes within social scenes , 2009, Vision Research.

[25]  Christof Koch,et al.  Predicting human gaze using low-level saliency combined with face detection , 2007, NIPS.

[26]  Xiongkuo Min,et al.  Brightness preserving video contrast enhancement using S-shaped Transfer function , 2013, 2013 Visual Communications and Image Processing (VCIP).

[27]  G. Alpers,et al.  Happy mouth and sad eyes: scanning emotional facial expressions. , 2011, Emotion.

[28]  Lihi Zelnik-Manor,et al.  Context-Aware Saliency Detection , 2012, IEEE Trans. Pattern Anal. Mach. Intell..

[29]  Xiongkuo Min,et al.  Influence of Spatial Resolution on State-of-the-Art Saliency Models , 2015, PCM.

[30]  Deva Ramanan,et al.  Face detection, pose estimation, and landmark localization in the wild , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[31]  Xianming Liu,et al.  Blind quality assessment of compressed images via pseudo structural similarity , 2016, 2016 IEEE International Conference on Multimedia and Expo (ICME).

[32]  Martin D. Levine,et al.  Visual Saliency Based on Scale-Space Analysis in the Frequency Domain , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[33]  Weisi Lin,et al.  Visual Saliency Detection With Free Energy Theory , 2015, IEEE Signal Processing Letters.

[34]  Liqing Zhang,et al.  Saliency Detection: A Spectral Residual Approach , 2007, 2007 IEEE Conference on Computer Vision and Pattern Recognition.

[35]  A. Coutrot,et al.  How saliency, faces, and sound influence gaze in dynamic social scenes. , 2014, Journal of vision.

[36]  Wenjun Zhang,et al.  Automatic Contrast Enhancement Technology With Saliency Preservation , 2015, IEEE Transactions on Circuits and Systems for Video Technology.

[37]  Weisi Lin,et al.  Saliency-based stereoscopic image retargeting , 2016, Inf. Sci..

[38]  Yu Fu,et al.  Visual saliency detection by spatially weighted dissimilarity , 2011, CVPR 2011.

[39]  Debopam Das,et al.  On the evolution of counter rotating vortex ring formed ahead of a compressible vortex ring , 2009, J. Vis..

[40]  Qi Zhao,et al.  Saliency in Crowd , 2014, ECCV.

[41]  Xiongkuo Min,et al.  Fixation prediction through multimodal analysis , 2015, 2015 Visual Communications and Image Processing (VCIP).

[42]  Weisi Lin,et al.  A Psychovisual Quality Metric in Free-Energy Principle , 2012, IEEE Transactions on Image Processing.

[43]  Ali Borji,et al.  Quantitative Analysis of Human-Model Agreement in Visual Saliency Modeling: A Comparative Study , 2013, IEEE Transactions on Image Processing.

[44]  John K. Tsotsos,et al.  Saliency Based on Information Maximization , 2005, NIPS.

[45]  J. Henderson,et al.  Do the eyes really have it? Dynamic allocation of attention when viewing moving faces. , 2012, Journal of vision.

[46]  Naila Murray,et al.  Saliency estimation using a non-parametric low-level vision model , 2011, CVPR 2011.

[47]  Zhou Wang,et al.  Unified Blind Quality Assessment of Compressed Natural, Graphic, and Screen Content Images , 2017, IEEE Transactions on Image Processing.

[48]  G. W. Snedecor Statistical Methods , 1964 .

[49]  Xiongkuo Min,et al.  Sound influences visual attention discriminately in videos , 2014, 2014 Sixth International Workshop on Quality of Multimedia Experience (QoMEX).

[50]  Xiongkuo Min,et al.  Fixation prediction through multimodal analysis , 2015, VCIP.

[51]  Chih-Jen Lin,et al.  LIBLINEAR: A Library for Large Linear Classification , 2008, J. Mach. Learn. Res..