Automatic Video Summarization by Spatio-temporal Analysis and Non-trivial Repeating Pattern Detection

Video content summarization provides an effective way to accelerating video browsing and retrieval. In this paper, we propose a novel approach to automatically generate the video summary. Firstly, the video structure is analyzed by spatio-temporal analysis. Then, we detect video non-trivial repeating patterns to remove the visual-content redundancy among video stream. Moreover, an importance evaluation model (IEM) is adopted to automatically determine the importance of each video shot according to the user need. This aims to construct video summarization with the most informative shots selected from groups of similar shots. Experimental results indicate that the proposed algorithm is more effective than existing approaches in video summarization generation.

[1]  Yihong Gong Summarizing Audiovisual Contents of a Video Program , 2003, EURASIP J. Adv. Signal Process..

[2]  Yueting Zhuang,et al.  Adaptive key frame extraction using unsupervised clustering , 1998, Proceedings 1998 International Conference on Image Processing. ICIP98 (Cat. No.98CB36269).

[3]  Thomas S. Huang,et al.  Constructing table-of-content for videos , 1999, Multimedia Systems.

[4]  Xiaoqing Ding,et al.  Real-time rotation invariant face detection based on cost-sensitive AdaBoost , 2003, Proceedings 2003 International Conference on Image Processing (Cat. No.03CH37429).

[5]  Arbee L. P. Chen,et al.  Efficient theme and non-trivial repeating pattern discovering in music databases , 1999, Proceedings 15th International Conference on Data Engineering (Cat. No.99CB36337).

[6]  Jitendra Malik,et al.  Normalized cuts and image segmentation , 1997, Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[7]  Yueting Zhuang,et al.  Key issues in video summarization and its application , 2003, Fourth International Conference on Information, Communications and Signal Processing, 2003 and the Fourth Pacific Rim Conference on Multimedia. Proceedings of the 2003 Joint.