University of Oulu’s MediaTeam research group participated to the manual and interactive search tasks with our video retrieval system, which is constructed out of three sub-applications: Content-based Query Tool, Cluster-Temporal Browser and Result Container with relevance feedback mechanism. The search engine in the heart of our search system creates result sets using three distinct search sub-engines that take care of the text, semantic concept and example based queries. The sub-applications and subengines result in varying search strategies. Therefore their effect on overall search performance has been studied in our experiments. The principal contribution in our search system is the Cluster-Temporal Browser, the effect of which we tested this year in conjunction with the traditional content-based paradigm: recurrent queries with relevance feedback. Seven search runs were submitted for the TRECVID 2005 search task: I_A_2_OUMT_I1Q_1: interactive with browsing disabled, expert users I_A_2_OUMT_I2B_2: interactive with browsing enabled, expert users I_A_2_OUMT_I3Q_3: interactive with browsing disabled, novice users I_A_2_OUMT_I4B_4: interactive with browsing enabled, novice users M_A_1_OUMT_M5T_5: manual with official text transcripts M_A_2_OUMT_M6TS_6: manual with official text transcripts and selected semantic concepts M_A_2_OUMT_M7TE_7: manual with official text transcripts and selected topic examples For interactive experiments, we have used eight test users: four novice and four expert users. In a carefully designed experimental setup, all users have used two system configurations: one with and one without Cluster-Temporal Browser. In our experiments we found that by incorporating Cluster-Temporal Browser the performance gain is more than 12% over the conventional content-based tools with novice users. We also found that on average expert users gained more than 18% better search performance over novice users, which shows that the test design has a significant effect to the outcome of the interactive test. Our manual experiment is a comparison of three query configurations: a baseline text only search, text combined with visual example search and text combined with semantic concept search. To facilitate a concept search, we trained a set of concept detectors (to name a few: newsroom, entertainment, maps and charts, news footage, weather, sports, outdoors). Our example search was based on low level descriptions (color and structure). The goal of the experiment was to find out how much more visual search examples and user defined semantic concepts contribute to the search performance over a text baseline. Our experimental results showed that the combined text and semantic concept search gives about 19% better performance over text baseline whereas text combined with example based search gives about 25% performance gain over the baseline. The results show that specific visual search examples accumulate better overall precision than the queries defined with our detected set of semantic concepts.
[1]
Sargur N. Srihari,et al.
Decision Combination in Multiple Classifier Systems
,
1994,
IEEE Trans. Pattern Anal. Mach. Intell..
[2]
Timo Ojala,et al.
Analysing the performance of visual, concept and text features in content-based video retrieval
,
2004,
MIR '04.
[3]
Joshua R. Smith,et al.
A Web-based System for Collaborative Annotation of Large Image and Video Collections
,
2005
.
[4]
Thorsten Joachims,et al.
Making large scale SVM learning practical
,
1998
.
[5]
James M. Rehg,et al.
Learning a Rare Event Detection Cascade by Direct Feature Selection
,
2003,
NIPS.
[6]
Timo Ojala,et al.
CONTENT-BASED BROWSING IN LARGE NEWS VIDEO DATABASES
,
2005
.
[7]
Timo Ojala,et al.
Cluster-temporal browsing of large news video databases
,
2004,
2004 IEEE International Conference on Multimedia and Expo (ICME) (IEEE Cat. No.04TH8763).
[8]
Hsuan-Tien Lin,et al.
A note on Platt’s probabilistic outputs for support vector machines
,
2007,
Machine Learning.
[9]
Christian Petersohn.
Fraunhofer HHI at TRECVID 2004: Shot Boundary Detection System
,
2004,
TRECVID.
[10]
Mika Rautiainen,et al.
Temporal color correlograms for video retrieval
,
2002,
Object recognition supported by user interaction for service robots.
[12]
Timo Ojala,et al.
Semantic image retrieval with hsv correlograms
,
2001
.
[13]
Mika Rautiainen,et al.
Detecting Semantic Concepts from Video Using Temporal Gradients and Audio Classification
,
2003,
CIVR.