A System for Video Recommendation using Visual Saliency, Crowdsourced and Automatic Annotations

In this paper we present a system for content-based video recommendation that exploits visual saliency to better represent video features and content\footnote{Demo video available at http://bit.ly/1FYloeQ}. Visual saliency is used to select relevant frames to be presented in a web-based interface to tag and annotate video frames in a social network; it is also employed to summarize video content to create a more effective video representation used in the recommender system. The system exploits automatic annotations from CNN-based classifiers on salient frames and user generated annotations. We evaluate several baseline approaches and show how the proposed method improves over them.

[1]  Junqing Yu,et al.  Affection arousal based highlight extraction for soccer video , 2013, Multimedia Tools and Applications.

[2]  Ian H. Witten,et al.  Learning to link with wikipedia , 2008, CIKM '08.

[3]  L. Itti,et al.  Modeling the influence of task on attention , 2005, Vision Research.

[4]  Yu He,et al.  The YouTube video recommendation system , 2010, RecSys '10.

[5]  Lixin Gao,et al.  The impact of YouTube recommendation system on video views , 2010, IMC '10.