Ontology driven contextual tagging of multimedia data

To exhibit multi-modal information and to facilitate people in finding multimedia resources, tagging plays a significant role. Various public events like protests and demonstrations are always consequences of break out of some public outrage resulting from prolonged exploitation and harassment. This outrage can be seen in news footage, blogs, text news and other web data. So, aggregating this variety of data from heterogeneous sources is a prerequisite step for tagging multimedia data with appropriate content. Since content has no meaning without a context, a video should be tagged with its relevant context and content information to assist user in multimedia retrieval. This paper proposes a model for tagging of multimedia data on the basis of contextual meaning. Since context is knowledge based, it has to be guided and learned by ontology which will help fragmented information to be represented in a more meaningful way. Our tagging approach is novel and has practical applicability in the sense that whenever a new video is uploaded on some media sharing site, the context and content information gets attached automatically to a video. Thus, providing relatively complete information associated with the video.

[1]  Meng Wang,et al.  Correlative multilabel video annotation with temporal kernels , 2008, TOMCCAP.

[2]  Bingbing Ni,et al.  Assistive tagging: A survey of multimedia tagging with human-computer joint exploration , 2012, CSUR.

[3]  David E. Millard,et al.  Artequakt: Generating Tailored Biographies with Automatically Annotated Fragments from the Web , 2002, SAAKM@ECAI.

[4]  Luciano Sbaiz,et al.  Finding meaning on YouTube: Tag recommendation and category discovery , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[5]  John R. Smith,et al.  Large-scale concept ontology for multimedia , 2006, IEEE MultiMedia.

[6]  Ilaria Bartolini,et al.  Shiatsu: semantic-based hierarchical automatic tagging of videos by segmentation using cuts , 2010, AIEMPro '10.

[7]  R. Manmatha,et al.  Multiple Bernoulli relevance models for image and video annotation , 2004, Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2004. CVPR 2004..

[8]  Meng Wang,et al.  Beyond Distance Measurement: Constructing Neighborhood Similarity for Video Annotation , 2009, IEEE Transactions on Multimedia.

[9]  John R. Smith,et al.  VideoAnnEx: IBM MPEG-7 Annotation Tool for Multimedia Indexing and Concept Learning , 2003 .

[10]  Santanu Chaudhury,et al.  Extending MOWL for Event Representation (E-MOWL) , 2013, 2013 IEEE/WIC/ACM International Joint Conferences on Web Intelligence (WI) and Intelligent Agent Technologies (IAT).

[11]  Tao Mei,et al.  Correlative multi-label video annotation , 2007, ACM Multimedia.

[12]  Gerard Salton,et al.  Research and Development in Information Retrieval , 1982, Lecture Notes in Computer Science.

[13]  Santanu Chaudhury,et al.  Ontology Specification and Integration for Multimedia Applications , 2007, Ontologies.