Learning to Boost Bottom-Up Fixation Prediction in Driving Environments via Random Forest

Saliency detection, an important step in many computer vision applications, can, for example, predict where drivers look in a vehicular traffic environment. While many bottom-up and top-down saliency detection models have been proposed for fixation prediction in outdoor scenes, no specific attempt has been made for traffic images. Here, we propose a learning saliency detection model based on a random forest (RF) to predict drivers’ fixation positions in a driving environment. First, we extract low-level (color, intensity, orientation, etc.) and high-level (e.g., the vanishing point and center bias) features and then predict the fixation points via RF-based learning. Finally, we evaluate the performance of our saliency prediction model qualitatively and quantitatively. We use quantitative evaluation metrics that include the revised receiver operating characteristic (ROC), the area under the ROC curve value, and the normalized scan-path saliency score. The experimental results on real traffic images indicate that our model can more accurately predict a driver’s fixation area, while driving than the state-of-the-art bottom-up saliency models.

[1]  Tao Deng,et al.  Top-down based saliency model in traffic driving environment , 2014, 17th International IEEE Conference on Intelligent Transportation Systems (ITSC).

[2]  Christof Koch,et al.  A Model of Saliency-Based Visual Attention for Rapid Scene Analysis , 2009 .

[3]  Asha Iyer,et al.  Components of bottom-up gaze allocation in natural images , 2005, Vision Research.

[4]  Jean Ponce,et al.  Vanishing point detection for road detection , 2009, CVPR.

[5]  Xin Gao,et al.  Segmentation-Based Salient Object Detection , 2015, CCCV.

[6]  Christof Koch,et al.  Learning a saliency map using fixated locations in natural scenes. , 2011, Journal of vision.

[7]  Benjamin W Tatler,et al.  The central fixation bias in scene viewing: selecting an optimal viewing position independently of motor biases and image feature distributions. , 2007, Journal of vision.

[8]  John K. Tsotsos,et al.  Saliency Based on Information Maximization , 2005, NIPS.

[9]  Huchuan Lu,et al.  Fixation prediction with a combined model of bottom-up saliency and vanishing point , 2015, 2016 IEEE Winter Conference on Applications of Computer Vision (WACV).

[10]  L. Itti,et al.  Modeling the influence of task on attention , 2005, Vision Research.

[11]  Giovanni Seni,et al.  Ensemble Methods in Data Mining: Improving Accuracy Through Combining Predictions , 2010, Ensemble Methods in Data Mining.

[12]  Tim K Marks,et al.  SUN: A Bayesian framework for saliency using natural statistics. , 2008, Journal of vision.

[13]  Ali Borji,et al.  What/Where to Look Next? Modeling Top-Down Visual Attention in Complex Interactive Environments , 2014, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[14]  Luis Salgado,et al.  Real-Time Vanishing Point Estimation in Road Sequences Using Adaptive Steerable Filter Banks , 2007, ACIVS.

[15]  Christopher Rasmussen Texture-Based Vanishing Point Voting for Road Shape Estimation , 2004, BMVC.

[16]  J. Henderson,et al.  Object-based attentional selection in scene viewing. , 2010, Journal of vision.

[17]  Hui Li,et al.  A Unified Framework for Salient Structure Detection by Contour-Guided Visual Search , 2015, IEEE Transactions on Image Processing.

[18]  Liqing Zhang,et al.  Saliency Detection: A Spectral Residual Approach , 2007, 2007 IEEE Conference on Computer Vision and Pattern Recognition.

[19]  Laurent Itti,et al.  Biologically Inspired Mobile Robot Vision Localization , 2009, IEEE Transactions on Robotics.

[20]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[21]  Tao Deng,et al.  Where Does the Driver Look? Top-Down-Based Saliency Detection in a Traffic Driving Environment , 2016, IEEE Transactions on Intelligent Transportation Systems.

[22]  David Crundall,et al.  Driver's visual attention as a function of driving experience and visibility. Using a driving simulator to explore drivers' eye movements in day, night and rain driving. , 2010, Accident; analysis and prevention.

[23]  Christof Koch,et al.  Modeling attention to salient proto-objects , 2006, Neural Networks.

[24]  C.-C. Jay Kuo,et al.  Learning a Combined Model of Visual Saliency for Fixation Prediction , 2016, IEEE Transactions on Image Processing.

[25]  W. Sardha Wijesoma,et al.  Fast Vanishing-Point Detection in Unstructured Environments , 2012, IEEE Transactions on Image Processing.

[26]  D. M. Green,et al.  Signal detection theory and psychophysics , 1966 .

[27]  Laurent Itti,et al.  Ieee Transactions on Pattern Analysis and Machine Intelligence 1 Rapid Biologically-inspired Scene Classification Using Features Shared with Visual Attention , 2022 .

[28]  Shuo Wang,et al.  Predicting human gaze beyond pixels. , 2014, Journal of vision.

[29]  Hui Kong,et al.  Generalizing Laplacian of Gaussian Filters for Vanishing-Point Detection , 2013, IEEE Transactions on Intelligent Transportation Systems.

[30]  M. Bindemann Scene and screen center bias early eye movements in scene viewing , 2010, Vision Research.

[31]  Pietro Perona,et al.  Graph-Based Visual Saliency , 2006, NIPS.

[32]  Yi Yang,et al.  Corner-surround Contrast for saliency detection , 2012, Proceedings of the 21st International Conference on Pattern Recognition (ICPR2012).

[33]  C. Koch,et al.  Computational modelling of visual attention , 2001, Nature Reviews Neuroscience.

[34]  Christopher M. Brown,et al.  Control of selective perception using bayes nets and decision theory , 1994, International Journal of Computer Vision.

[35]  Laurent Itti,et al.  Beyond bottom-up: Incorporating task-dependent influences into a computational model of spatial attention , 2007, 2007 IEEE Conference on Computer Vision and Pattern Recognition.

[36]  Frédo Durand,et al.  Learning to predict where humans look , 2009, 2009 IEEE 12th International Conference on Computer Vision.

[37]  Hong Yan,et al.  Bayes Saliency-Based Object Proposal Generator for Nighttime Traffic Images , 2018, IEEE Transactions on Intelligent Transportation Systems.