Effective multimedia surveillance using a human-centric approach

Large-scale multimedia surveillance installations usually consist of a number of spatially distributed video cameras that are installed in a premise and are connected to a central control station, where human operators (e.g., security personnel) remotely monitor the scene images captured by the cameras. In the majority of these systems the ratio of human operators to the number of camera views is very low. This potentially raises the problem that some important events may be missed. Studies have shown that a human operator can effectively monitor only four camera views. Moreover, the visual attention of human operator drops below the acceptable level while performing the task of visual monitoring. Therefore, there is a need for the selection of the four most relevant camera views at a given time instant. This paper proposes a human-centric approach to solve the problem of dynamically selecting and scheduling the four best camera views. In the proposed approach we use a feedback camera to observe the human monitoring the surveillance camera feeds. Using this information, the system computes the operator’s attention to the camera views to automatically determine the importance of events being captured by the respective cameras. This real-time non-invasive relevance feedback is then augmented with the automatic detection of events to compute the four best feeds. The experiments show the effectiveness of the proposed approach by improving the identification of important events occurring in the environment.

[1]  Verónica Vilaplana,et al.  Region-based mean shift tracking: Application to face tracking , 2008, 2008 15th IEEE International Conference on Image Processing.

[2]  Ramesh Jain,et al.  Experiential Sampling for video surveillance , 2003, IWVS '03.

[3]  A. Hampapur,et al.  Smart video surveillance: exploring the concept of multiscale spatiotemporal tracking , 2005, IEEE Signal Processing Magazine.

[4]  Marc Davis Active capture: integrating human-computer interaction and computer vision/audition to automate media capture , 2003, 2003 International Conference on Multimedia and Expo. ICME '03. Proceedings (Cat. No.03TH8698).

[5]  Ramesh C. Jain,et al.  ACM SIGMM retreat report on future directions in multimedia research , 2005, TOMCCAP.

[6]  Badrinath Roysam,et al.  Image change detection algorithms: a systematic survey , 2005, IEEE Transactions on Image Processing.

[7]  Mohan M. Trivedi,et al.  An eye localization, tracking and blink pattern recognition system: Algorithm and evaluation , 2010, TOMCCAP.

[8]  Yongdong Zhang,et al.  Human attention model for semantic scene analysis in movies , 2008, 2008 IEEE International Conference on Multimedia and Expo.

[9]  C. Koch,et al.  Computational modelling of visual attention , 2001, Nature Reviews Neuroscience.

[10]  Mohan S. Kankanhalli,et al.  Information assimilation framework for event detection in multimedia surveillance systems , 2006, Multimedia Systems.

[11]  Pradeep K. Atrey,et al.  A hierarchical model for representation of events in multimedia observation systems , 2009, EiMM '09.

[12]  Amarnag Subramanya,et al.  Real time eye tracking for human computer interfaces , 2003, 2003 International Conference on Multimedia and Expo. ICME '03. Proceedings (Cat. No.03TH8698).

[13]  Pierre Baldi,et al.  Bayesian surprise attracts human attention , 2005, Vision Research.

[14]  Kostas Karpouzis,et al.  Estimation of behavioral user state based on eye gaze and head pose—application in an e-learning environment , 2009, Multimedia Tools and Applications.

[15]  Pradeep K. Atrey,et al.  Modeling and assessing quality of information in multisensor multimedia monitoring systems , 2011, TOMCCAP.

[16]  Mubarak Shah,et al.  Monitoring head/eye motion for driver alertness with one camera , 2000, Proceedings 15th International Conference on Pattern Recognition. ICPR-2000.

[17]  J.G. Taylor,et al.  Modelling human attention and emotions , 2004, 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No.04CH37541).

[18]  Christopher E. Peters,et al.  Attention-driven eye gaze and blinking for virtual humans , 2003, SIGGRAPH '03.

[19]  Karon E. MacLean,et al.  Emulating human attention-getting practices with wearable haptics , 2010, 2010 IEEE Haptics Symposium.

[20]  Yusuf Sinan Akgül,et al.  Eye-gaze based real-time surveillance video synopsis , 2009, Pattern Recognit. Lett..

[21]  Wenjun Chris Zhang,et al.  Human attention modeling in a human-machine interface based on the incorporation of contextual features in a Bayesian network , 2005, 2005 IEEE International Conference on Systems, Man and Cybernetics.

[22]  Riad I. Hammoud,et al.  Real-time estimation of human attention field in LWIR and color surveillance videos , 2008, 2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops.

[23]  M.J.T. Reinders,et al.  Eye Tracking by Template Matching using an Automatic Codebook Generation Scheme , 2004 .

[24]  Mohan S. Kankanhalli,et al.  Finding Interesting Images in Albums using Attention , 2008, J. Multim..

[25]  W. Eric L. Grimson,et al.  Adaptive background mixture models for real-time tracking , 1999, Proceedings. 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149).

[26]  Zafer Savas,et al.  REAL-TIME DETECTION AND TRACKING OF HUMAN EYES IN VIDEO , 2005 .

[27]  Pradeep K. Atrey,et al.  Automatic scheduling of CCTV camera views using a human-centric approach , 2008, 2008 IEEE International Conference on Multimedia and Expo.

[28]  Jorge Dias,et al.  Face tracking based on haar-like features and eigenfaces , 2004 .

[29]  Sergio A. Velastin,et al.  How close are we to solving the problem of automated visual surveillance? , 2008, Machine Vision and Applications.