The COST292 experimental framework for rushes summarization task in TRECVID 2008

In this paper, the method used for Rushes Summarization task by the COST 292 consortium is reported. The approach proposed this year differs significantly from the one proposed in the previous years because of the introduction of new processing steps, like repetition detection in scenes. The method starts with junk frames removal and follows with clustering and scene detection; then for each scene, repetitions are detected in order to extract once the real scene; the following step consists in face detections (faces are considered semantically relevant) and in pan, tilt and zoom detections (other camera motions are usually related to technical operations in the backstage); finally the summary is extracted.

[1]  Akio Yamada,et al.  The MPEG-7 color layout descriptor: a compact image feature description for high-speed image/video segment retrieval , 2001, Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205).

[2]  Xueming Qian,et al.  Effective Fades and Flashlight Detection Based on Accumulating Histogram Difference , 2006, IEEE Transactions on Circuits and Systems for Video Technology.

[3]  Jianbo Shi,et al.  Learning Segmentation by Random Walks , 2000, NIPS.

[4]  Jitendra Malik,et al.  Normalized cuts and image segmentation , 1997, Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[5]  Jenny Benois-Pineau,et al.  Detection of visual dialog scenes in video content based on structural and semantic features , 2005 .

[6]  Paul Over,et al.  The trecvid 2008 BBC rushes summarization evaluation , 2008, TVS '08.

[7]  Jianbo Shi,et al.  A Random Walks View of Spectral Segmentation , 2001, AISTATS.