Film Comic Generation with Eye Tracking

Automatic generation of film comic requires solving several challenging problems such as selecting important frames well conveying the whole story, trimming the frames to fit the shape of panels without corrupting the composition of original image and arranging visually pleasing speech balloons without hiding important objects in the panel. We propose a novel approach to the automatic generation of film comic. The key idea is to aggregate eye-tracking data and image features into a computational map, called iMap, for quantitatively measuring the importance of frames in terms of story content and user attention. The transition of iMap in time sequences provides the solution to frame selection. Word balloon arrangement and image trimming are realized as the results of optimizing the energy functions derived from the iMap.

[1]  Majid Mirmehdi,et al.  Video Indexing using Motion Estimation , 2003, BMVC.

[2]  Jörn Loviscach,et al.  From movie to comic, informed by the screenplay , 2007, SIGGRAPH '07.

[3]  Degang Zhang,et al.  Human-centered attention models for video summarization , 2010, ICMI-MLMI '10.

[4]  Hwan-Gue Cho,et al.  Cinema comics: Cartoon generation from video stream , 2006, GRAPP.

[5]  Guizhong Liu,et al.  A Multiple Visual Models Based Perceptive Analysis Framework for Multilevel Video Summarization , 2007, IEEE Transactions on Circuits and Systems for Video Technology.

[6]  Xian-Sheng Hua,et al.  Bridging low-level features and high-level semantics via fMRI brain imaging for video classification , 2010, ACM Multimedia.

[7]  Sarah V. Porter,et al.  Video Segmentation and Indexing using Motion Estimation , 2004 .

[8]  Masahiro Toyoura,et al.  Film Comic Reflecting Camera-Works , 2012, MMM.

[9]  Jhing-Fa Wang,et al.  A Novel Video Summarization Based on Mining the Story-Structure and Semantic Relations Among Concept Entities , 2009, IEEE Transactions on Multimedia.

[10]  Wei-Ta Chu,et al.  A User Experience Model for Home Video Summarization , 2009, MMM.

[11]  Anan Liu,et al.  Watching, Thinking, Reacting: A Human-Centered Framework for Movie Content Analysis , 2010, J. Digit. Content Technol. its Appl..

[12]  Meng Wang,et al.  Movie2Comics: a feast of multimedia artwork , 2010, ACM Multimedia.

[13]  Matthieu Guillaumin,et al.  Combining Image-Level and Segment-Level Models for Automatic Annotation , 2012, MMM.

[14]  Luc Van Gool,et al.  Speeded-Up Robust Features (SURF) , 2008, Comput. Vis. Image Underst..

[15]  Harry W. Agius,et al.  Video summarisation: A conceptual framework and survey of the state of the art , 2008, J. Vis. Commun. Image Represent..

[16]  Lie Lu,et al.  A generic framework of user attention model and its application in video summarization , 2005, IEEE Trans. Multim..

[17]  John Lee,et al.  Eye movements and pupil dilation during event perception , 2006, ETRA '06.

[18]  Shingo Uchihashi,et al.  An interactive comic book presentation for exploring video , 2000, CHI.