ARO: Exploring the Design of Smart-Ring Interactions for Encumbered Hands

Fingertip computing has seen increased interest through miniaturized smart-rings for augmenting digital peripherals. One key advantages of such always-available input devices is the non-necessity to hold a device for interaction, as it remains affixed to a finger for access when needed. Such a wearable device makes it possible to interaction with content even when the hand is encumbered, by grasping or holding objects. Our investigation aims at understanding the properties of this fundamental smart-ring advantage. We designed a smart-ring prototype, ARO (in-Air, on-Ring, on-Object interaction), which facilitates input while grasping objects. To better identify interaction possibilities, we present the results of an elicitation study through which we grouped various forms of micro-gestures possible with ARO while holding objects under different grasp requirements. We then explored the ability for users to perform different navigation tasks (i.e. zooming and panning) using the smart-ring with encumbered hands. In our studies, users were most efficient when using either In-air or On-ring interactions, in comparisons to gestures detected On-object. Furthermore, In-air was the most preferred by our participants. Based on our findings, we conclude with recommendations for the design of future smart-rings and fingertip devices at large, to allow efficient interaction while hands are encumbered.

[1]  Yang Li,et al.  Tap, swipe, or move: attentional demands for distracted smartphone input , 2012, AVI.

[2]  Sean White,et al.  Nenya: subtle and eyes-free mobile input with a magnetically-tracked finger ring , 2011, CHI.

[3]  Suranga Nanayakkara,et al.  ChewIt. An Intraoral Interface for Discreet Interactions , 2019, CHI.

[4]  Jürgen Steimle,et al.  DeformWear , 2017, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[5]  Bongwon Suh,et al.  OctaRing: Examining Pressure-Sensitive Multi-Touch Input on a Finger Ring Device , 2016, UIST.

[6]  Zheng Li,et al.  How Subtle Can It Get? , 2020, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[7]  Suranga Nanayakkara,et al.  Digital Digits , 2015, ACM Comput. Surv..

[8]  Marcos Serrano,et al.  Exploring the use of hand-to-face input for interacting with head-worn displays , 2014, CHI.

[9]  L. Besançon,et al.  An interaction Continuum for 3D Dataset Visualization , 2017 .

[10]  Laurence Nigay,et al.  A Framework for the Combination and Characterization of Output Modalities , 2000, DSV-IS.

[11]  Daniel Vogel,et al.  RotoSwype: Word-Gesture Typing using a Ring , 2019, CHI.

[12]  Shahin Tajik,et al.  Multi-sensor Based Gestures Recognition with a Smart Finger Ring , 2014, HCI.

[13]  Sang Ho Yoon,et al.  Wearable textile input device with multimodal sensing for eyes-free mobile interaction during daily activities , 2016, Pervasive Mob. Comput..

[14]  Suranga Nanayakkara,et al.  Thumb-In-Motion: Evaluating Thumb-to-Ring Microgestures for Athletic Activity , 2018, SUI.

[15]  J. L. Roux An Introduction to the Kalman Filter , 2003 .

[16]  Sahin Albayrak,et al.  eRing: multiple finger gesture recognition with one ring using an electric field , 2015, iWOAR.

[17]  Pierre Dragicevic,et al.  Fair Statistical Communication in HCI , 2016 .

[18]  Gregory D. Abowd,et al.  FingerSound , 2017, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[19]  Xing-Dong Yang,et al.  Magic finger: always-available input through finger instrumentation , 2012, UIST.

[20]  Ki-Woong Park,et al.  A Ubiquitous Fashionable Computer with an i-Throw Device on a Location-Based Service Environment , 2007, 21st International Conference on Advanced Information Networking and Applications Workshops (AINAW'07).

[21]  Taylor Cl,et al.  The anatomy and mechanics of the human hand. , 1955, Artificial limbs.

[22]  Jürgen Steimle,et al.  Grasping Microgestures: Eliciting Single-hand Microgestures for Handheld Objects , 2019, CHI.

[23]  Claudio Pacchierotti,et al.  A three DoFs wearable tactile display for exploration and manipulation of virtual objects , 2012, 2012 IEEE Haptics Symposium (HAPTICS).

[24]  Radu-Daniel Vatavu,et al.  Formalizing Agreement Analysis for Elicitation Studies: New Measures, Significance Test, and Toolkit , 2015, CHI.

[25]  Yoon Sang Kim,et al.  A new wearable input device: SCURRY , 2005, IEEE Transactions on Industrial Electronics.

[26]  Meredith Ringel Morris,et al.  User-defined gestures for surface computing , 2009, CHI.

[27]  Teddy Seyed,et al.  User Elicitation on Single-hand Microgestures , 2016, CHI.

[28]  Radu-Daniel Vatavu,et al.  From Controls on the Steering Wheel to Controls on the Finger: Using Smart Rings for In-Vehicle Interactions , 2020, Conference on Designing Interactive Systems.

[29]  Ken Hinckley,et al.  LightRing: always-available 2D input on any surface , 2014, UIST.

[30]  Pourang Irani,et al.  D-SWIME: A Design Space for Smartwatch Interaction Techniques Supporting Mobility and Encumbrance , 2018, CHI.

[31]  Pierre Dragicevic,et al.  Running an HCI experiment in multiple parallel universes , 2014, CHI Extended Abstracts.

[32]  Hirotaka Osawa,et al.  iRing: intelligent ring using infrared reflection , 2012, UIST.

[33]  Anne Roudaut,et al.  Understanding Grip Shifts: How Form Factors Impact Hand Movements on Mobile Phones , 2017, CHI.

[34]  Teng Han,et al.  Combining Ring Input with Hand Tracking for Precise, Natural Interaction with Spatial Analytic Interfaces , 2016, SUI.

[35]  Jean Vanderdonckt,et al.  Gestures for Smart Rings: Empirical Results, Insights, and Design Implications , 2018, Conference on Designing Interactive Systems.

[36]  Yoshinobu Tonomura,et al.  “Body coupled FingerRing”: wireless wearable keyboard , 1997, CHI.

[37]  K. Tsukada,et al.  UBI-FINGER: A SIMPLE GESTURE INPUT DEVICE FOR MOBILE AND UBIQUITOUS ENVIRONMENT , 2003 .

[38]  Yiqiang Chen,et al.  ThermalRing: Gesture and Tag Inputs Enabled by a Thermal Imaging Smart Ring , 2020, CHI.

[39]  Geehyuk Lee,et al.  Voice+Tactile: Augmenting In-vehicle Voice User Interface with Tactile Touchpad Interaction , 2020, CHI.

[40]  Pourang Irani,et al.  ThumbText: Text Entry for Wearable Devices Using a Miniature Ring , 2018, Graphics Interface.

[41]  Ian Oakley,et al.  Nailz: Sensing Hand Input with Touch Sensitive Nails , 2020, CHI.

[42]  Kyu-Han Kim,et al.  TypingRing: A Wearable Ring Platform for Text Input , 2015, MobiSys.

[43]  Daniel Vogel,et al.  WRIST: Watch-Ring Interaction and Sensing Technique for Wrist Gestures and Macro-Micro Pointing , 2019, MobileHCI.

[44]  Sean White,et al.  uTrack: 3D input using two magnetic sensors , 2013, UIST.

[45]  Dhairya Dand,et al.  A Simple Wearable Hand Gesture Recognition Device Using iMEMS , 2009, 2009 International Conference of Soft Computing and Pattern Recognition.

[46]  Joseph A. Paradiso,et al.  NailO: Fingernails as an Input Surface , 2015, CHI.

[47]  Bongshin Lee,et al.  Reducing legacy bias in gesture elicitation studies , 2014, INTR.

[48]  Dongyang Zhao,et al.  BiTipText: Bimanual Eyes-Free Text Entry on a Fingertip Keyboard , 2020, CHI.

[49]  M. Baker Statisticians issue warning over misuse of P values , 2016, Nature.

[50]  Wen-Huang Cheng,et al.  FingerPad: private and subtle interaction using fingertips , 2013, UIST.

[51]  Yui-Pan Yau How Subtle Can It Get? A Trimodal Study of Ring-sized Interfaces for One-Handed Drone Control , 2020 .

[52]  Sang Ho Yoon,et al.  Plex: finger-worn textile sensor for mobile interaction during activities , 2014, UbiComp Adjunct.

[53]  Sang Ho Yoon,et al.  BikeGesture: User Elicitation and Performance of Micro Hand Gesture as Input for Cycling , 2017, CHI Extended Abstracts.

[54]  Sang Ho Yoon,et al.  TIMMi: Finger-worn Textile Input Device with Multimodal Sensing in Mobile Interaction , 2015, TEI.

[55]  Joanna Bergstrom-Lehtovirta,et al.  Modeling the functional area of the thumb on mobile touchscreen surfaces , 2014, CHI.

[56]  Hyoukryeol Choi,et al.  Development of Soft-Actuator-Based Wearable Tactile Display , 2008, IEEE Transactions on Robotics.

[57]  Eric Lecolinet,et al.  Clutch-free panning and integrated pan-zoom control on touch-sensitive surfaces: the cyclostar approach , 2010, CHI.

[58]  Michael Rohs,et al.  A Taxonomy of Microinteractions: Defining Microgestures Based on Ergonomic and Scenario-Dependent Requirements , 2011, INTERACT.

[59]  Junbo Wang,et al.  Magic Ring: a self-contained gesture input device on finger , 2013, MUM.

[60]  José Luis Pons Rovira,et al.  The MANUS-HAND Dextrous Robotics Upper Limb Prosthesis: Mechanical and Manipulation Aspects , 2004, Auton. Robots.

[61]  Yiqiang Chen,et al.  A ring-shaped interactive device for large remote display and mobile device control , 2011, UbiComp '11.

[62]  Jean Vanderdonckt,et al.  A Systematic Review of Gesture Elicitation Studies: What Can We Learn from 216 Studies? , 2020, Conference on Designing Interactive Systems.

[63]  Heni Ben Amor,et al.  Grasp Recognition for Uncalibrated Data Gloves: A Machine Learning Approach , 2008, PRESENCE: Teleoperators and Virtual Environments.

[64]  Chris Harrison,et al.  Abracadabra: wireless, high-precision, and unpowered finger input for very small mobile devices , 2009, UIST '09.

[65]  Stephen A. Brewster,et al.  The Effects of Encumbrance and Mobility on Touch-Based Gesture Interactions for Mobile Phones , 2015, MobileHCI.

[66]  G. Schlesinger Der mechanische Aufbau der künstlichen Glieder , 1919 .

[67]  Yasuhito Suenaga,et al.  “FingeRing”: a full-time wearable interface , 1994, CHI '94.

[68]  Seongil Lee,et al.  Design of an Integrated Wearable Multimedia Interface for In-Vehicle Telematics , 2004, PCM.