Managing the video content for searching and summarizing has become a challenging task. Extracting semantics from video scenes enables information to be presented in a more understandable manner. Finding the semantics between video contexts is a difficult task; much recent research has focused on this issue. Most videos, such as TV serials and commercial movies, are character- centric. Therefore, the context and relationship between characters needs to be organized systematically to analyze the video. So, it is necessary to identify the contextual relationships between characters in the scene and the video. We propose Character-Net, a network structure. It finds characters in a group of shots, extracts the speaker and listeners in the scene, represents it with character-based graphs and draws the relationship between all characters by accumulating the character-based graphs at video. In this paper, we describe how to build Character-Net. Experimental results show Character-Net is an effective methodology to extract the major characters in videos.
[1]
Volker Roth.
Content-based retrieval from digital video
,
1999,
Image Vis. Comput..
[2]
Timothy F. Cootes,et al.
Feature Detection and Tracking with Constrained Local Models
,
2006,
BMVC.
[3]
Andrew Zisserman,et al.
Taking the bite out of automated naming of characters in TV video
,
2009,
Image Vis. Comput..
[4]
Rong Yan,et al.
Multiple instance learning for labeling faces in broadcasting news video
,
2005,
MULTIMEDIA '05.
[5]
Ebroul Izquierdo,et al.
Logotype detection to support semantic-based video annotation
,
2007,
Signal Process. Image Commun..
[6]
Li Li,et al.
Semantic event representation and recognition using syntactic attribute graph grammar
,
2009,
Pattern Recognit. Lett..