Adapting smartwatch interfaces to hand gestures during movements: offset models and the C-shaped pattern of tapping

Interacting with smartwatches is fairly common when users are moving. Although novel interaction gestures like flick of the wrist are implemented, basic touch gestures such as tapping and swiping still dominate. Using these gestures during a variety of movements could be challenging, and it is still not clear how the interface of smartwatches should tailor to users’ gestures during movements and how the usage of smartwatches influences the pattern of users’ movements. Therefore, this study investigates the interrelationship among users’ interaction gestures, movements, and gait features. An experiment was conducted among 47 participants, who used smartwatches through tapping, swiping, and wrist flicking to complete daily tasks in stand, strolling, normal walking, rushing, and jogging. They were tracked through built-in accelerometer and angle sensors. Four findings were derived from the experiment. First, rushing and jogging significantly decrease the effectiveness and efficiency of tapping. To reduce the tapping deviation, offset models were proposed and tested. Second, there is a C-shaped pattern on the round screen where tapping targets achieves higher accuracy than other areas. Third, the tapping performance could be improved by setting target sizes. Target sizes at 0.7 cm in stand, 1.1 cm in strolling, and 1.1 cm in walking achieve a high level of accuracy (95%), while target sizes at 1.5 cm in rushing and jogging achieve a middle level of accuracy (90%). Finally, tapping, swiping, and wrist flicking when users are moving significantly reduce their gait symmetry and step length. They do not imply significant influence on gait intensity, regularity, and overall stability.

[1]  Sean Hayes,et al.  Tablet Interaction Accuracy and Precision , 2014 .

[2]  Lior Rokach,et al.  User identity verification via mouse dynamics , 2012, Inf. Sci..

[3]  Mark H. Chignell,et al.  Mobile text entry: relationship between walking speed and text input task difficulty , 2005, Mobile HCI.

[4]  Shuping Yi,et al.  Adapting the Navigation Interface of Smart Watches to User Movements , 2017, Int. J. Hum. Comput. Interact..

[5]  Yang Li,et al.  Experimental analysis of touch-screen gesture designs in mobile environments , 2011, CHI.

[6]  Chris Harrison,et al.  Abracadabra: wireless, high-precision, and unpowered finger input for very small mobile devices , 2009, UIST '09.

[7]  Elena Karahanna,et al.  Time Flies When You're Having Fun: Cognitive Absorption and Beliefs About Information Technology Usage , 2000, MIS Q..

[8]  Vivian Genaro Motti,et al.  Micro interactions and Multi dimensional Graphical User Interfaces in the Design of Wrist Worn Wearables , 2015 .

[9]  Kent Lyons,et al.  An investigation into round touchscreen wristwatch interaction , 2008, Mobile HCI.

[10]  James R. Lewis Psychometric Evaluation of the PSSUQ Using Data from Five Years of Usability Studies , 2002 .

[11]  Stephen A. Brewster,et al.  The Effects of Encumbrance and Mobility on Touch-Based Gesture Interactions for Mobile Phones , 2015, MobileHCI.

[12]  D. Damos Multiple-task performance , 2020 .

[13]  Ian Oakley,et al.  TriTap: Identifying Finger Touches on Smartwatches , 2017, CHI.

[14]  Gierad Laput,et al.  Expanding the input expressivity of smartwatches with mechanical pan, twist, tilt and click , 2014, CHI.

[15]  Buntarou Shizuki,et al.  B2B-Swipe: Swipe Gesture for Rectangular Smartwatches from a Bezel to a Bezel , 2016, CHI.

[16]  RauPei-Luen Patrick,et al.  Playing with multiple wearable devices , 2015 .

[17]  Jorunn L Helbostad,et al.  Estimation of gait cycle characteristics by trunk accelerometry. , 2004, Journal of biomechanics.

[18]  Ian Oakley,et al.  Beats: Tapping Gestures for Smart Watches , 2015, CHI.

[19]  Stephen A. Brewster,et al.  Comparing evaluation methods for encumbrance and walking on interaction with touchscreen mobile devices , 2014, MobileHCI '14.

[20]  Ian Oakley,et al.  The Flat Finger: Exploring Area Touches on Smartwatches , 2016, CHI.

[21]  Enrico Rukzio,et al.  Investigating selection and reading performance on a mobile phone while walking , 2010, Mobile HCI.

[22]  Barbara S. Chaparro,et al.  Is Touch-Based Text Input Practical for a Smartwatch? , 2015, HCI.

[23]  Gregory D. Abowd,et al.  WatchOut: extending interactions on a smartwatch with inertial sensing , 2016, SEMWEB.

[24]  Ji Jung Lim,et al.  Visual search on a mobile device while walking , 2012, Mobile HCI.

[25]  Pei-Luen Patrick Rau,et al.  Exploring factors influencing multitasking interaction with multiple smart devices , 2013, Comput. Hum. Behav..

[26]  Pei-Luen Patrick Rau,et al.  Playing with multiple wearable devices: Exploring the influence of display, motion and gender , 2015, Comput. Hum. Behav..

[27]  P. Rau,et al.  Application of smart bracelet to monitor frailty‐related gait parameters of older Chinese adults: A preliminary study , 2018, Geriatrics & gerontology international.

[28]  Ravin Balakrishnan,et al.  PageFlip: Leveraging Page-Flipping Gestures for Efficient Command and Value Selection on Smartwatches , 2018, CHI.

[29]  R. Fitzpatrick,et al.  Acceleration patterns of the head and pelvis when walking on level and irregular surfaces. , 2003, Gait & posture.

[30]  G. Muthumari,et al.  Mouse gesture based authentication using machine learning algorithm , 2014, 2014 IEEE International Conference on Advanced Communications, Control and Computing Technologies.

[31]  Markus Löchtefeld,et al.  A user-specific machine learning approach for improving touch accuracy on mobile devices , 2012, UIST '12.

[32]  Thomas Alexander,et al.  Analysis of Visual Performance during the Use of Mobile Devices While Walking , 2014, HCI.

[33]  Andrew Sears,et al.  An empirical comparison of use-in-motion evaluation scenarios for mobile computing devices , 2005, Int. J. Hum. Comput. Stud..

[34]  Daniel Buschek,et al.  User-specific touch models in a cross-device context , 2013, MobileHCI '13.

[35]  E. G. Cruz,et al.  Stereotypical fingertip trajectories during grasp. , 2003, Journal of neurophysiology.

[36]  S. Sueda,et al.  How the lumbrical muscle contributes to placing the fingertip in space: a three dimensional cadaveric study to assess fingertip trajectory and metacarpophalangeal joint balancing , 2016, The Journal of hand surgery, European volume.

[37]  James R. Lewis,et al.  Psychometric Evaluation of the PSSUQ Using Data from Five Years of Usability Studies , 2002, Int. J. Hum. Comput. Interact..

[38]  Yvonne Rogers,et al.  Fat Finger Worries: How Older and Younger Users Physically Interact with PDAs , 2005, INTERACT.

[39]  Jacob O. Wobbrock,et al.  Getting off the treadmill: evaluating walking user interfaces for mobile devices in public spaces , 2008, Mobile HCI.

[40]  Brian Caulfield,et al.  SHIMMER: A new tool for temporal gait analysis , 2009, 2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society.

[41]  Andrew Sears,et al.  How do people tap when walking? An empirical investigation of nomadic data entry , 2007, Int. J. Hum. Comput. Stud..

[42]  Vivian Genaro Motti,et al.  Smart Wearables or Dumb Wearables?: Understanding how Context Impacts the UX in Wrist Worn Interaction , 2016, SIGDOC.

[43]  Barry A. T. Brown,et al.  Situating Wearables: Smartwatch Use in Context , 2017, CHI.

[44]  Jacob O. Wobbrock,et al.  WalkType: using accelerometer data to accomodate situational impairments in mobile touch screen text entry , 2012, CHI.

[45]  Thomas Alexander,et al.  Optimal Touch Button Size for the use of Mobile Devices while Walking , 2015 .

[46]  F. L. van Noort,et al.  Effect of running on throughput in pointing tasks: a Fitts' law experiment , 2015 .

[47]  Carl Gutwin,et al.  Faster Command Selection on Touchscreen Watches , 2016, CHI.

[48]  Roderick Murray-Smith,et al.  Nomadic Input on Mobile Devices: The Influence of Touch Input Technique and Walking Speed on Performance and Offset Modeling , 2016, Hum. Comput. Interact..

[49]  Liwen Wang,et al.  Chinese-Based Spearcons: Improving Pedestrian Navigation Performance in Eyes-Free Environment , 2016, Int. J. Hum. Comput. Interact..