Lecture video segmentation by automatically analyzing the synchronized slides

In this paper we propose a solution which segments lecture video by analyzing its supplementary synchronized slides. The slides content derives automatically from OCR (Optical Character Recognition) process with an approximate accuracy of 90%. Then we partition the slides into different subtopics by examining their logical relevance. Since the slides are synchronized with the video stream, the subtopics of the slides indicate exactly the segments of the video. Our evaluation reveals that the average length of segments for each lecture is ranged from 5 to 15 minutes, and 45% segments achieved from test datasets are logically reasonable.

[1]  Irena Koprinska,et al.  Temporal video segmentation: A survey , 2001, Signal Process. Image Commun..

[2]  Raúl Rojas,et al.  Anthropocentric Video Segmentation for Lecture Webcasts , 2008, EURASIP J. Image Video Process..

[3]  Harald Sack,et al.  Lecture Video Indexing and Analysis Using Video OCR Technology , 2011, 2011 Seventh International Conference on Signal Image Technology & Internet-Based Systems.

[4]  Kunio Fukunaga,et al.  Blackboard segmentation using video image of lecture and its applications , 2000, Proceedings 15th International Conference on Pattern Recognition. ICPR-2000.

[5]  Faouzi Alaya Cheikh,et al.  Semantic Tags for Lecture Videos , 2012, 2012 IEEE Sixth International Conference on Semantic Computing.

[6]  Ignas Budvytis,et al.  Semi-Supervised Video Segmentation Using Tree Structured Graphical Models , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[7]  Jay F. Nunamaker,et al.  Segmentation of lecture videos based on text: a method combining multiple linguistic features , 2004, 37th Annual Hawaii International Conference on System Sciences, 2004. Proceedings of the.

[8]  Yasuo Ariki,et al.  Topic segmentation and retrieval system for lecture videos based on spontaneous speech recognition , 2003, INTERSPEECH.