Increasing the expressive power of gesture-based interaction on mobile devices. (Augmenter le pouvoir d'expression de l'interaction gestuelle sur les appareils mobiles)

Current mobile interfaces let users directly manipulate the objects displayed on the screen with simple stroke gestures, e.g. tap on soft buttons or menus or pinch to zoom. To access a larger command space, the users are often forced to go through long steps, making the interaction cumbersome and inefficient. More complex gestures offer a powerful way to access information quickly and to perform a command more efficiently [5]. However, they are more difficult to learn and control. Gesture typing [78] is an interesting alternative to input text: it lets users draw a gesture on soft keyboards to enter text, from the first until the final letter in a word. In this thesis, I increase the expressive power of mobile interaction by leveraging the gesture’s shape and dynamics and the screen space to produce rich output, to invoke commands, and to facilitate appropriation in different contexts of use. I design "Expressive Keyboard" that transforms the gesture variations into rich output, and demonstrate several applications in the context of textbased communication. As well, I propose "CommandBoard", a gesture keyboard that lets users efficiently select commands from a large command space while supporting the transition from novices to experts. I demonstrate different applications of "CommandBoard", each offers users a choice, based on their cognitive and motor skills, as well as the size and organization of the current command set. Altogether, these techniques give users more expressive power by leveraging human’s motor control and cognitive ability to learn, to control, and to appropriate.

[1]  Kasper Hornbæk,et al.  Vulture: a mid-air word-gesture keyboard , 2014, CHI.

[2]  William Buxton,et al.  The limits of expert performance using hierarchic marking menus , 1993, INTERCHI.

[3]  Shumin Zhai,et al.  Differences and Similarities between Finger and Pen Stroke Gestures on Stationary and Mobile devices , 2015, TCHI.

[4]  Geehyuk Lee,et al.  Evaluation of a Smart-Restorable Backspace Technique to Facilitate Text Entry Error Correction , 2016, CHI.

[5]  Shumin Zhai,et al.  The performance of touch screen soft buttons , 2009, CHI.

[6]  Per Ola Kristensson,et al.  Memorability of pre-designed and user-defined gesture sets , 2013, CHI.

[7]  P. Fitts The information capacity of the human motor system in controlling the amplitude of movement. , 1954, Journal of experimental psychology.

[8]  Miles MacLeod,et al.  Usability measurement in context , 1994, Behav. Inf. Technol..

[9]  Shumin Zhai,et al.  FFitts law: modeling finger touch with fitts' law , 2013, CHI.

[10]  Marcelo M. Wanderley,et al.  The Importance of Parameter Mapping in Electronic Instrument Design , 2002, NIME.

[11]  Tovi Grossman,et al.  A survey of software learnability: metrics, methodologies and guidelines , 2009, CHI.

[12]  Shumin Zhai,et al.  Using strokes as command shortcuts: cognitive benefits and toolkit support , 2009, CHI.

[13]  Shumin Zhai,et al.  The word-gesture keyboard: reimagining keyboard interaction , 2012, CACM.

[14]  Klaus H. Hinrichs,et al.  Modeling Human Performance of Stroke-Based Text Entry , 2015, MobileHCI.

[15]  Vittorio Fuccella,et al.  Gestures and widgets: performance in text editing on multi-touch capable mobile devices , 2013, CHI.

[16]  Bonnie A. Nardi,et al.  An ethnographic study of distributed problem solving in spreadsheet development , 1990, CSCW '90.

[17]  James A. Landay,et al.  Visual similarity of pen gestures , 2000, CHI.

[18]  William Buxton,et al.  Chunking and Phrasing and the Design of Human-Computer Dialogues (Invited Paper) , 1995, IFIP Congress.

[19]  Shumin Zhai,et al.  Performance and User Experience of Touchscreen and Gesture Keyboards in a Lab Setting and in the Wild , 2015, CHI.

[20]  Wendy E. Mackay,et al.  Beyond Grids, Interactive Graphical Substrates to Structure Digital Layout , 2018 .

[21]  Patrick Baudisch,et al.  Understanding touch , 2011, CHI.

[22]  Shumin Zhai,et al.  “Writing with music”: Exploring the use of auditory feedback in gesture interfaces , 2008, TAP.

[23]  Patrick Baudisch,et al.  The generalized perceived input point model and how to double touch accuracy by extracting fingerprints , 2010, CHI.

[24]  Wendy E. Mackay,et al.  Fieldward and Pathward: Dynamic Guides for Defining Your Own Gestures , 2017, CHI.

[25]  Olivier Chapuis,et al.  Controlling widgets with one power-up button , 2013, UIST.

[26]  Robert Pastel,et al.  Measuring the difficulty of steering through corners , 2006, CHI.

[27]  I. Scott MacKenzie,et al.  Phrase sets for evaluating text entry techniques , 2003, CHI Extended Abstracts.

[28]  Emmanuel Pietriga,et al.  Side pressure for bidirectional navigation on small devices , 2013, MobileHCI '13.

[29]  Niels Taatgen,et al.  Toward a unified theory of the multitasking continuum: from concurrent performance to task switching, interruption, and resumption , 2009, CHI.

[30]  William A. Woods What's Important About Knowledge Representation? , 1983, Computer.

[31]  Anastasia Bezerianos,et al.  Coordination of tilt and touch in one- and two-handed use , 2014, CHI.

[32]  Wendy E. Mackay,et al.  Interactive paper substrates to support musical creation , 2012, CHI.

[33]  Geehyuk Lee,et al.  TapBoard 2: Simple and Effective Touchpad-like Interaction on a Multi-Touch Surface Keyboard , 2016, CHI.

[34]  Wendy E. Mackay,et al.  Expressive Keyboards: Enriching Gesture-Typing on Mobile Devices , 2016, UIST.

[35]  M. Sheelagh T. Carpendale,et al.  KeyStrokes: Personalizing Typed Text with Visualization , 2007, EuroVis.

[36]  Atau Tanaka,et al.  Beyond recognition: using gesture variation for continuous interaction , 2013, CHI Extended Abstracts.

[37]  Shumin Zhai,et al.  Interlaced QWERTY: accommodating ease of visual search and input flexibility in shape writing , 2008, CHI.

[38]  Barry A. T. Brown,et al.  100 days of iPhone use: understanding the details of mobile device use , 2014, MobileHCI '14.

[39]  Shumin Zhai,et al.  Octopus: evaluating touchscreen keyboard correction and recognition algorithms via , 2013, CHI.

[40]  Xiaojun Bi,et al.  CommandBoard: Creating a General-Purpose Command Gesture Input Space for Soft Keyboard , 2017, UIST.

[41]  Matthias Felleisen,et al.  On the Expressive Power of Programming Languages , 1990, European Symposium on Programming.

[42]  Shumin Zhai,et al.  Modeling human performance of pen stroke gestures , 2007, CHI.

[43]  Mark S. Pfaff,et al.  Designing multi-touch gestures to support emotional expression in IM , 2014, CHI Extended Abstracts.

[44]  Shumin Zhai,et al.  Command strokes with and without preview: using pen gestures on keyboard for command selection , 2007, CHI.

[45]  James R. Eagan,et al.  Augmented letters: mnemonic gesture-based shortcuts , 2013, CHI.

[46]  Keith A. Butler,et al.  Connecting theory and practice: a case study of achieving usability goals , 1985, CHI '85.

[47]  Shumin Zhai,et al.  Shorthand writing on stylus keyboard , 2003, CHI '03.

[48]  W. Mackay Responding to cognitive overload : Co-adaptation between users and technology , 2000 .

[49]  Olivier Bau,et al.  OctoPocus: a dynamic guide for learning gesture-based command sets , 2008, UIST '08.

[50]  Joonhwan Lee,et al.  Smiley face: why we use emoticon stickers in mobile messaging , 2016, MobileHCI Adjunct.

[51]  Alan J. Dix,et al.  Designing for appropriation , 2007, BCS HCI.

[52]  Olivier Bau,et al.  Arpège: learning multitouch chord gestures vocabularies , 2013, ITS.

[53]  Terry Winograd,et al.  FlowMenu: combining command, text, and data entry , 2000, UIST '00.

[54]  Yang Li Gesture search: a tool for fast mobile data access , 2010, UIST '10.

[55]  Shumin Zhai,et al.  Hard lessons: effort-inducing interfaces benefit spatial learning , 2007, CHI.

[56]  Shumin Zhai,et al.  Touch behavior with different postures on soft smartphone keyboards , 2012, Mobile HCI.

[57]  Florian Alt,et al.  There is more to Typing than Speed: Expressive Mobile Touch Keyboards via Dynamic Font Personalisation , 2015, MobileHCI.

[58]  Jun Rekimoto,et al.  Expressive typing: a new way to sense typing pressure and its applications , 2009, CHI Extended Abstracts.

[59]  David Goldberg,et al.  Touch-typing with a stylus , 1993, INTERCHI.

[60]  Shumin Zhai,et al.  Foundational Issues in Touch-Surface Stroke Gesture Design - An Integrative Review , 2012, Found. Trends Hum. Comput. Interact..

[61]  Wendy E. Mackay,et al.  Structured observation with polyphony: a multifaceted tool for studying music composition , 2014, Conference on Designing Interactive Systems.

[62]  Shumin Zhai,et al.  SHARK2: a large vocabulary shorthand writing system for pen-based computers , 2004, UIST '04.

[63]  Tim Paek,et al.  Usability guided key-target resizing for soft keyboards , 2010, IUI '10.

[64]  Wendy E. Mackay,et al.  Which interaction technique works when?: floating palettes, marking menus and toolglasses support different task strategies , 2002, AVI '02.

[65]  Albert N. Badre,et al.  Discount Learnability Evaluation , 1995 .

[66]  Wendy E. Mackay,et al.  Generative walkthroughs: to support creative redesign , 2009, C&C '09.

[67]  Craig S. Miller,et al.  Comparison of Mouse and Keyboard Efficiency , 2010 .

[68]  Jonathan Bean,et al.  Learning from IKEA hacking: i'm not one to decoupage a tabletop and call it a day. , 2009, CHI.

[69]  Atau Tanaka,et al.  Understanding Gesture Expressivity through Muscle Sensing , 2015, ACM Trans. Comput. Hum. Interact..