Gesturing in the Wild : Understanding the Effects and Implications of Gesture-Based Interaction for Dynamic Presentations

Driven by the increasing availability of low-cost sensing hardware, gesture-based input is quickly becoming a viable form of interaction for a variety of applications. Electronic presentations (e.g., PowerPoint, Keynote) have long been seen as a natural fit for this form of interaction. However, despite 20 years of prototyping such systems, little is known about how gesture-based input affects presentation dynamics, or how it can be best applied in this context. Instead, past work has focused almost exclusively on recognition algorithms. This paper explicitly addresses these gaps in the literature. Through observations of real-world practices, we first describe the types of gestures presenters naturally make and the purposes these gestures serve when presenting content. We then introduce Maestro, a gesture-based presentation system explicitly designed to support and enhance these existing practices. Finally, we describe the results of a real-world field study in which Maestro was evaluated in a classroom setting for several weeks. Our results indicate that gestures which enable direct interaction with slide content are the most natural fit for this input modality. In contrast, we found that using gestures to navigate slides (the most common implementation in all prior systems) has significant drawbacks. Our results also show how gesture-based input can noticeably alter presentation dynamics, often in ways that are not desirable.

[1]  Joel Lanir,et al.  Observing presenters' use of visual aids to inform the design of classroom presentation software , 2008, CHI.

[2]  John S. Boreczky,et al.  Manipulating and Annotating Slides in a Multi-Display Environment , 2003, INTERACT.

[3]  Gregory Dudek,et al.  A natural gesture interface for operating robotic systems , 2008, 2008 IEEE International Conference on Robotics and Automation.

[4]  Michael J. Black,et al.  Recognizing temporal trajectories using the condensation algorithm , 1998, Proceedings Third IEEE International Conference on Automatic Face and Gesture Recognition.

[5]  Anthony Tang,et al.  Shadow reaching: a new perspective on interaction for large displays , 2007, UIST.

[6]  Rahul Sukthankar,et al.  Self-Calibrating Camera-Assisted Presentation Interface , 2000 .

[7]  François Bérard,et al.  Bare-hand human-computer interaction , 2001, PUI '01.

[8]  Xiang Cao,et al.  Evaluation of alternative presentation control techniques , 2005, CHI EA '05.

[9]  Jin-Hyung Kim,et al.  An HMM-Based Threshold Model Approach for Gesture Recognition , 1999, IEEE Trans. Pattern Anal. Mach. Intell..

[10]  Keechul Jung,et al.  Recognition-based gesture spotting in video games , 2004, Pattern Recognit. Lett..

[11]  David Salesin,et al.  On creating animated presentations , 2003, SCA '03.

[12]  Jerry Alan Fails,et al.  Light widgets: interacting in every-day spaces , 2002, IUI '02.

[13]  Desney S. Tan,et al.  Pre-emptive shadows: eliminating the blinding light from projectors , 2002, CHI Extended Abstracts.

[14]  Lijun Tang,et al.  A portable system for anywhere interactions , 2004, CHI EA '04.

[15]  Michel Beaudouin-Lafon,et al.  Charade: remote control of objects using free-hand gestures , 1993, CACM.