A Multidisciplinary Approach to Unlocking Television Broadcast Archives

Abstract Audiovisual material is a vital component of the world's heritage but it remains difficult to access. With the Netherlands Institute for Sound and Vision as one of its partners, the MuNCH project aims to investigate new methods for improving access to a wide range of audiovisual documents. MuNCH brings together three research fields: multimedia analysis, language technology and semantic technologies. Within the MuNCH project, we have investigated several combinations of these fields. We have compared text matching, ontology querying, and semantic visual querying as methods to translate a multimedia query to the vocabulary of the retrieval system. In addition, we have investigated how users make such a translation, and have used this as a benchmark to create automatic methods. We have used multimedia technology to automatically detect objects and scenes as they occur in video, and made use of language technology to exploit automatic transcriptions of speech. We have enriched the Sound and Vision thesaurus that is used to annotate the TV programmes in order to provide a user with a wider range of search results. In order to verify the results of the project against real user needs, MuNCH has participated in the creation of a logging system which monitors the usage of the Sound and Vision catalogue system. Insights in the needs of real users will be used as input for all three of MuNCH's research strands.

[1]  Colin Potts,et al.  Design of Everyday Things , 1988 .

[2]  Ellen M. Voorhees,et al.  Query expansion using lexical-semantic relations , 1994, SIGIR '94.

[3]  Vladimir Cherkassky,et al.  The Nature Of Statistical Learning Theory , 1997, IEEE Trans. Neural Networks.

[4]  Marcel Worring,et al.  Content-Based Image Retrieval at the End of the Early Years , 2000, IEEE Trans. Pattern Anal. Mach. Intell..

[5]  Vladimir N. Vapnik,et al.  The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.

[6]  Christiane Fellbaum,et al.  Book Reviews: WordNet: An Electronic Lexical Database , 1999, CL.

[7]  R. Edmondson Audiovisual archiving : philosophy and principles , 2004 .

[8]  Jarkko Kari,et al.  Facing and bridging gaps in Web searching , 2006, Inf. Process. Manag..

[9]  Dennis Koelma,et al.  The MediaMill TRECVID 2008 Semantic Video Search Engine , 2008, TRECVID.

[10]  Marcel Worring,et al.  The Mediamill Semantic Video Search Engine , 2007, 2007 IEEE International Conference on Acoustics, Speech and Signal Processing - ICASSP '07.

[11]  Bob J. Wielinga,et al.  Patterns of semantic relations to improve image content search , 2007, J. Web Semant..

[12]  Andrew Tomkins,et al.  Guest Editors' Introduction: Social Media and Search , 2007, IEEE Internet Computing.

[13]  Maarten de Rijke,et al.  Exploiting redundancy in cross-channel video retrieval , 2007, MIR '07.

[14]  Véronique Malaisé,et al.  Enriching a Thesaurus to Improve Retrieval of Audiovisual Documents , 2008, SAMT.

[15]  Katja Hofmann,et al.  Assessing concept selection for video retrieval , 2008, MIR '08.

[16]  Véronique Malaisé,et al.  Automatic Annotation Suggestions for Audiovisual Archives: Evaluation Aspects , 2009 .

[17]  Ynze van Houten Searching for videos : the structure of video interaction in the framework of information foraging theory , 2009 .

[18]  Huajun Chen,et al.  The Semantic Web , 2011, Lecture Notes in Computer Science.