Belly gestures: body centric gestures on the abdomen

Recent HCI research has shown that the body offers an interactive surface particularly suitable to eyes-free interaction. While researchers have mainly focused on the arms and the hands, we argue that the surface of the belly is especially appropriate. The belly offers a fairly large surface that can be easily reached with the two hands in any circumstance, including walking or running. We report on a study that explored how users perform one-handed gestures on their abdomen. Users use different mental spatial orientations depending on the complexity of the gesture they have to draw (drawing a digit vs. a simple directional stroke). When provided with no visual orientation cues they often draw gestures following symmetries relative to a horizontal or vertical axis. The more complex the gesture, the less stability in orientation. Focusing on directional strokes, we found that users are able to draw almost linear gestures, despite the fact that the abdomen is not perfectly planar, and perform particularly well in cardinal directions. The paper ends up with some guidelines that may inform the design of novel interaction techniques.

[1]  Patrick Baudisch,et al.  Imaginary phone: learning imaginary interfaces by transferring spatial memory from a familiar device , 2011, UIST.

[2]  Patrick Baudisch,et al.  Touch input on curved surfaces , 2011, CHI.

[3]  Chris Harrison,et al.  Whack gestures: inexact and inattentive interaction with mobile devices , 2010, TEI '10.

[4]  Kasper Hornbæk,et al.  An experimental comparison of touch interaction on vertical and horizontal surfaces , 2012, NordiCHI.

[5]  C. Spence,et al.  The science of interpersonal touch: An overview , 2010, Neuroscience & Biobehavioral Reviews.

[6]  Jan O. Borchers,et al.  Pinstripe: eyes-free continuous input on interactive clothing , 2011, CHI.

[7]  Gabriel Baud-Bovy,et al.  The haptic perception of spatial orientations , 2008, Experimental Brain Research.

[8]  Takashi Maeno,et al.  Touch interface on back of the hand , 2011, SIGGRAPH '11.

[9]  L M Parsons,et al.  Perceived spatial organization of cutaneous patterns on surfaces of the human body in various positions. , 1987, Journal of experimental psychology. Human perception and performance.

[10]  Meredith Ringel Morris,et al.  User-defined gestures for surface computing , 2009, CHI.

[11]  Stephen A. Brewster,et al.  Usable gestures for mobile interfaces: evaluating social acceptability , 2010, CHI.

[12]  Chris Harrison,et al.  OmniTouch: wearable multitouch interaction everywhere , 2011, UIST.

[13]  Xing-Dong Yang,et al.  Magic finger: always-available input through finger instrumentation , 2012, UIST.

[14]  Bing-Yu Chen,et al.  Pub - point upon body: exploring eyes-free interaction and methods on an arm , 2011, UIST.

[15]  Daniel J. Wigdor,et al.  Direct-touch vs. mouse input for tabletop displays , 2007, CHI.

[16]  F. Royer,et al.  Normative data for the Symbol Digit Substitution Task. , 1981, Journal of clinical psychology.

[17]  Thomas Rosemeier,et al.  Interaction of vestibular, somatosensory and visual signals for postural control and motion perception under terrestrial and microgravity conditions—a conceptual model , 1998, Brain Research Reviews.

[18]  William Buxton,et al.  Issues in combining marking and direct manipulation techniques , 1991, UIST '91.

[19]  Brad A. Myers,et al.  Maximizing the guessability of symbolic input , 2005, CHI Extended Abstracts.

[20]  W. A. Hodge,et al.  Trunk kinematics during locomotor activities. , 1992, Physical therapy.

[21]  Michael Rohs,et al.  ShoeSense: a new perspective on gestural interaction and wearable applications , 2012, CHI.

[22]  Lei Wang,et al.  Understanding flicking on curved surfaces , 2012, CHI.

[23]  Ivan Poupyrev,et al.  Touché: enhancing touch interaction on humans, screens, liquids, and everyday objects , 2012, CHI.

[24]  Virpi Roto,et al.  Interaction in 4-second bursts: the fragmented nature of attentional resources in mobile HCI , 2005, CHI.

[25]  P. Haggard,et al.  Touch and the body , 2010, Neuroscience & Biobehavioral Reviews.

[26]  S. Gandevia,et al.  The proprioceptive senses: their roles in signaling body shape, body position and movement, and muscle force. , 2012, Physiological reviews.

[27]  Richard Martin,et al.  Design for wearability , 1998, Digest of Papers. Second International Symposium on Wearable Computers (Cat. No.98EX215).

[28]  Desney S. Tan,et al.  Skinput: appropriating the body as an input surface , 2010, CHI.

[29]  Daniel Ashbrook Enabling mobile microinteractions , 2010 .

[30]  Wendy E. Mackay,et al.  Body-centric design space for multi-surface interaction , 2013, CHI.