MixT: automatic generation of step-by-step mixed media tutorials

Users of complex software applications often learn concepts and skills through step-by-step tutorials. Today, these tutorials are published in two dominant forms: static tutorials composed of images and text that are easy to scan, but cannot effectively describe dynamic interactions; and video tutorials that show all manipulations in detail, but are hard to navigate. We hypothesize that a mixed tutorial with static instructions and per-step videos can combine the benefits of both formats. We describe a comparative study of static, video, and mixed image manipulation tutorials with 12 participants and distill design guidelines for mixed tutorials. We present MixT, a system that automatically generates step-by-step mixed media tutorials from user demonstrations. MixT segments screencapture video into steps using logs of application commands and input events, applies video compositing techniques to focus on salient infor-mation, and highlights interactions through mouse trails. Informal evaluation suggests that automatically generated mixed media tutorials were as effective in helping users complete tasks as tutorials that were created manually.

[1]  Takeo Igarashi,et al.  Generating photo manipulation tutorials by demonstration , 2009, ACM Trans. Graph..

[2]  Susan Palmiter,et al.  Animated Demonstrations vs Written Instructions for Learning Procedural Tasks: A Preliminary Investigation , 1991, Int. J. Man Mach. Stud..

[3]  Tovi Grossman,et al.  Searching for software learning resources using application context , 2011, UIST.

[4]  Tovi Grossman,et al.  Sketch-sketch revolution: an engaging tutorial system for guided sketching and application learning , 2011, UIST.

[5]  Susan M. Harrison A comparison of still, animated, or nonillustrated on-line help with written or spoken instructions in a graphical user interface , 1995, CHI '95.

[6]  Mira Dontcheva,et al.  A Framework for content-adaptive photo manipulation macros: Application to face, landscape, and global manipulations , 2011, TOGS.

[7]  Tovi Grossman,et al.  ToolClips: an investigation of contextual video assistance for functionality understanding , 2010, CHI.

[8]  Tovi Grossman,et al.  CommunityCommands: command recommendations for software applications , 2009, UIST '09.

[9]  Björn Hartmann,et al.  Delta: a tool for representing and comparing workflows , 2012, CHI.

[10]  Rob Miller,et al.  Sikuli: using GUI screenshots for search and automation , 2009, UIST '09.

[11]  Caitlin Kelleher,et al.  Stencils-based tutorials: design and evaluation , 2005, CHI.

[12]  Maneesh Agrawala,et al.  ShowMeHow: Translating User Interface Instructions Between Similar Applications , 2011 .

[13]  Tovi Grossman,et al.  IP-QAT: in-product questions, answers, & tips , 2011, UIST.

[14]  Morgan Dixon,et al.  Prefab: implementing advanced behaviors using pixel-based reverse engineering of interface structure , 2010, CHI.

[15]  Björn Hartmann,et al.  ShowMeHow: translating user interface instructions between applications , 2011, UIST.

[16]  G. W. Furnas,et al.  Generalized fisheye views , 1986, CHI '86.

[17]  Roberto Brunelli,et al.  Template Matching Techniques in Computer Vision: Theory and Practice , 2009 .

[18]  Fabio Pellacini,et al.  MeshFlow : interactive visualization of mesh construction sequences , 2011, SIGGRAPH 2011.

[19]  Tessa A. Lau,et al.  DocWizards: a system for authoring follow-me documentation wizards , 2005, UIST.

[20]  Tovi Grossman,et al.  Chronicle: capture, exploration, and playback of document workflow histories , 2010, UIST.

[21]  Mira Dontcheva,et al.  Pause-and-play: automatically linking screencast video tutorials with applications , 2011, UIST.

[22]  Susan Palmiter,et al.  Animated Demonstrations for Learning Procedural Computer-Based Tasks , 1993, Hum. Comput. Interact..