Predicting Movie Trailer Viewer's “Like/Dislike” via Learned Shot Editing Patterns

Nowadays, there are many movie trailers publicly available on social media website such as YouTube, and many thousands of users have independently indicated whether they like or dislike those trailers. Although it is understandable that there are multiple factors that could influence viewers' like or dislike of the trailer, we aim to address a preference question in this work: Can subjective multimedia features be developed to predict the viewer's preference presented by like (by thumbs-up) or dislike (by thumbs-down) during and after watching movie trailers? We designed and implemented a computational framework that is composed of low-level multimedia feature extraction, feature screening and selection, and classification, and applied it to a collection of 725 movie trailers. Experimental results demonstrated that, among dozens of multimedia features, the single low-level multimedia feature of shot length variance is highly predictive of a viewer's “like/dislike” for a large portion of movie trailers. We interpret these findings such that variable shot lengths in a trailer tend to produce a rhythm that is likely to stimulate a viewer's positive preference. This conclusion was also proved by the repeatability experiments results using another 600 trailer videos and it was further interpreted by viewers'eye-tracking data.

[1]  Phuoc Tran-Gia,et al.  Quantification of YouTube QoE via Crowdsourcing , 2011, 2011 IEEE International Symposium on Multimedia.

[2]  Jaap Ham,et al.  Brightness differences influence the evaluation of affective pictures , 2013, Cognition & emotion.

[3]  Nicu Sebe,et al.  Exploiting facial expressions for affective video summarisation , 2009, CIVR '09.

[4]  Nilesh V. Patel,et al.  Video shot detection and characterization for video databases , 1997, Pattern Recognit..

[5]  Franziska Frankfurter,et al.  Film An International History Of The Medium , 2016 .

[6]  B. Manav Color‐emotion associations and color preferences: A case study for residences , 2007 .

[7]  R. A. Mcfarland Relationship of skin temperature changes to the emotions accompanying music , 1985, Biofeedback and self-regulation.

[8]  David Dagan Feng,et al.  Realistic Human Action Recognition with Audio Context , 2010, 2010 International Conference on Digital Image Computing: Techniques and Applications.

[9]  Nuria Oliver,et al.  Towards Computational Models of the Visual Aesthetic Appeal of Consumer Videos , 2010, ECCV.

[10]  Yo-Sub Han,et al.  Evaluation of User Reputation on YouTube , 2009, HCI.

[11]  Shiliang Zhang,et al.  Affective MTV analysis based on arousal and valence features , 2008, 2008 IEEE International Conference on Multimedia and Expo.

[12]  Ananda S. Chowdhury,et al.  Video key frame extraction through dynamic Delaunay clustering with a structural constraint , 2013, J. Vis. Commun. Image Represent..

[13]  M. Verboord,et al.  Dimensions of Conventionality and Innovation in Film: The Cultural Classification of Blockbusters, Award Winners, and Critics’ Favourites , 2014 .

[14]  Mohammad Soleymani,et al.  A Bayesian framework for video affective representation , 2009, 2009 3rd International Conference on Affective Computing and Intelligent Interaction and Workshops.

[15]  B. Fredrickson The role of positive emotions in positive psychology. The broaden-and-build theory of positive emotions. , 2001, The American psychologist.

[16]  S. Domnic,et al.  Shot based keyframe extraction for ecological video indexing and retrieval , 2014, Ecol. Informatics.

[17]  M. Hemphill,et al.  A note on adults' color-emotion associations. , 1996, The Journal of genetic psychology.

[18]  Shih-Fu Chang,et al.  Clustering methods for video browsing and annotation , 1996, Electronic Imaging.

[19]  J. Cutting,et al.  Attention and the Evolution of Hollywood Film , 2010, Psychological science.

[20]  Paul Over,et al.  Video shot boundary detection: Seven years of TRECVid activity , 2010, Comput. Vis. Image Underst..

[21]  Yeong-Ho Ha,et al.  Spatial color descriptor for image retrieval and video segmentation , 2003, IEEE Trans. Multim..

[22]  Michael J. Swain,et al.  Color indexing , 1991, International Journal of Computer Vision.

[23]  B. Detenber,et al.  A Bio‐Informational Theory of Emotion: Motion and Image Size Effects on Viewers , 1996 .

[24]  Ling-Yu Duan,et al.  Hierarchical movie affective content analysis based on arousal and valence features , 2008, ACM Multimedia.

[25]  J. Zlatev,et al.  Moving Ourselves, Moving Others. Motion and emotion in intersubjectivity, consciousness and language , 2012 .

[26]  K. H. Kim,et al.  Emotion recognition system using short-term monitoring of physiological signals , 2004, Medical and Biological Engineering and Computing.

[27]  José San Pedro,et al.  Ranking and classifying attractiveness of photos in folksonomies , 2009, WWW '09.

[28]  Keith M. Johnston,et al.  'The Coolest Way to Watch Movie Trailers in the World' , 2008 .

[29]  Greg Mori,et al.  Action recognition by learning mid-level motion features , 2008, 2008 IEEE Conference on Computer Vision and Pattern Recognition.

[30]  Richard Neupert A History of the French New Wave Cinema , 2002 .

[31]  W. Wirth,et al.  Media and Emotions , 2005 .

[32]  Boon-Lock Yeo,et al.  Rapid scene analysis on compressed video , 1995, IEEE Trans. Circuits Syst. Video Technol..

[33]  Christopher B. Stapleton,et al.  Mixed Reality and Experiential Movie Trailers: Combining Emotions and Immersion to Innovate Entertainment Marketing , 2005 .

[34]  Loong Fah Cheong,et al.  Affective understanding in film , 2006, IEEE Trans. Circuits Syst. Video Technol..

[35]  Xiaoou Tang,et al.  Photo and Video Quality Evaluation: Focusing on the Subject , 2008, ECCV.

[36]  Yong Shi,et al.  Fast Video Shot Boundary Detection Based on SVD and Pattern Matching , 2013, IEEE Transactions on Image Processing.

[37]  Slobodan Petrovic,et al.  Improving Effectiveness of Intrusion Detection by Correlation Feature Selection , 2010, 2010 International Conference on Availability, Reliability and Security.

[38]  Greta Hsu Jacks of All Trades and Masters of None: Audiences' Reactions to Spanning Genres in Feature Film Production , 2006 .

[39]  Elisabeth André,et al.  Emotion recognition based on physiological changes in music listening , 2008, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[40]  Kiyoharu Aizawa,et al.  Affective Audio-Visual Words and Latent Topic Driving Model for Realizing Movie Affective Scene Classification , 2010, IEEE Transactions on Multimedia.

[41]  Peter Y. K. Cheung,et al.  Affective Level Video Segmentation by Utilizing the Pleasure-Arousal-Dominance Information , 2008, IEEE Transactions on Multimedia.

[42]  Steven Albert Movie Stars and the Distribution of Financially Successful Films in the Motion Picture Industry , 1998 .

[43]  J. G. Taylor,et al.  Emotion recognition in human-computer interaction , 2005, Neural Networks.

[44]  W. Walls,et al.  Uncertainty in the Movie Industry: Does Star Power Reduce the Terror of the Box Office? , 1999 .

[45]  Luís Alberto da Silva Cruz,et al.  Retinal image quality assessment using generic image quality indicators , 2014, Inf. Fusion.

[46]  Cheng-Yu Hsieh,et al.  Efficient video segment matching for detecting temporal-based video copies , 2013, Neurocomputing.

[47]  Thierry Pun,et al.  Multimodal Emotion Recognition in Response to Videos , 2012, IEEE Transactions on Affective Computing.

[48]  Hyoungkwan Kim,et al.  Using Hue, Saturation, and Value Color Space for Hydraulic Excavator Idle Time Analysis , 2007 .

[49]  J. Gross,et al.  Emotion elicitation using films , 1995 .

[50]  C. Boyatzis,et al.  Children's emotional associations with colors. , 1994, The Journal of genetic psychology.

[51]  A. Schaefer,et al.  Please Scroll down for Article Cognition & Emotion Assessing the Effectiveness of a Large Database of Emotion-eliciting Films: a New Tool for Emotion Researchers , 2022 .

[52]  M. Bradley,et al.  Looking at pictures: affective, facial, visceral, and behavioral reactions. , 1993, Psychophysiology.

[53]  Shigeki Sagayama,et al.  Rhythm and Tempo Analysis Toward Automatic Music Transcription , 2007, 2007 IEEE International Conference on Acoustics, Speech and Signal Processing - ICASSP '07.

[54]  Mike Graham,et al.  Extracting information about emotions in films , 2003, ACM Multimedia.

[55]  Alan Hanjalic,et al.  Affective video content representation and modeling , 2005, IEEE Transactions on Multimedia.

[56]  Andreas Dengel,et al.  Attentive documents: Eye tracking as implicit feedback for information retrieval and beyond , 2012, TIIS.

[57]  Yuan Yan Tang,et al.  Recognizing complex events in real movies by combining audio and video features , 2014, Neurocomputing.

[58]  Susan T. Dumais,et al.  The good, the bad, and the random: an eye-tracking study of ad quality in web search , 2010, SIGIR.

[59]  Fabrice Souvannavong,et al.  Latent semantic analysis for an effective region-based video shot retrieval system , 2004, MIR '04.

[60]  Sabine Süsstrunk,et al.  Measuring colorfulness in natural images , 2003, IS&T/SPIE Electronic Imaging.

[61]  Yanqing Zhang,et al.  SVMs Modeling for Highly Imbalanced Classification , 2009, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[62]  R. Sklar Film: An International History of the Medium , 1993 .

[63]  Chang-Hsing Lee,et al.  Scene-based event detection for baseball videos , 2007, J. Vis. Commun. Image Represent..

[64]  Julie Delon,et al.  Movie and video scale-time equalization application to flicker reduction , 2006, IEEE Transactions on Image Processing.

[65]  Wolfgang Nejdl,et al.  How useful are your comments?: analyzing and predicting youtube comments and comment ratings , 2010, WWW '10.