3D Hand Pose Estimation on Conventional Capacitive Touchscreens

Contemporary mobile devices with touchscreens capture the X/Y position of finger tips on the screen and pass these coordinates to applications as though the input were points in space. Of course, human hands are much more sophisticated, able to form rich 3D poses capable of far more complex interactions than poking at a screen. In this paper, we describe how conventional capacitive touchscreens can be used to estimate 3D hand pose, enabling richer interaction opportunities. Importantly, our software-only approach requires no special or new sensors, either internal or external. As a proof of concept, we use an off-the-shelf Samsung Tablet flashed with a custom kernel. After describing our software pipeline, we report findings from our user study, we conclude with several example applications we built to illustrate the potential of our approach.

[1]  Niels Henze,et al.  Estimating the Finger Orientation on Capacitive Touchscreens Using Convolutional Neural Networks , 2017, ISS.

[2]  Mike Wu,et al.  A study of hand shape use in tabletop gesture interaction , 2006, CHI Extended Abstracts.

[3]  Mike Wu,et al.  Gesture registration, relaxation, and reuse for multi-point direct-touch surfaces , 2006, First IEEE International Workshop on Horizontal Interactive Human-Computer Systems (TABLETOP '06).

[4]  Chris Harrison,et al.  Super-Resolution Capacitive Touchscreens , 2021, CHI.

[5]  Lale Akarun,et al.  Real time hand pose estimation using depth sensors , 2011, 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops).

[6]  Eric Lecolinet,et al.  MicroRolls: expanding touch-screen input vocabulary by distinguishing rolls vs. slides of the thumb , 2009, CHI.

[7]  Simon Rogers,et al.  AnglePose: robust, precise capacitive touch tracking via 3d orientation estimation , 2011, CHI.

[8]  Antti Oulasvirta,et al.  How We Type: Movement Strategies and Performance in Everyday Typing , 2016, CHI.

[9]  Patrick Olivier,et al.  Digits: freehand 3D interactions anywhere using a wrist-worn gloveless sensor , 2012, UIST.

[10]  Xiang Cao,et al.  Detecting and leveraging finger orientation for interaction with direct-touch surfaces , 2009, UIST '09.

[11]  Haibo Li,et al.  Direct hand pose estimation for immersive gestural interaction , 2015, Pattern Recognit. Lett..

[12]  Yang Zhang,et al.  Tomo: Wearable, Low-Cost Electrical Impedance Tomography for Hand Gesture Recognition , 2015, UIST.

[13]  Xiang 'Anthony' Chen,et al.  The fat thumb: using the thumb's contact size for single-handed mobile interaction , 2012, Mobile HCI.

[14]  Daniel J. Wigdor,et al.  Rock & rails: extending multi-touch interactions with shape gestures to enable precise spatial manipulations , 2011, CHI.

[15]  Fabrice Matulic,et al.  Hand Contact Shape Recognition for Posture-Based Tabletop Widgets and Interaction , 2017, ISS.

[16]  Niels Henze,et al.  PalmTouch: Using the Palm as an Additional Input Modality on Commodity Smartphones , 2018, CHI.

[17]  Karan Singh,et al.  Eurographics/siggraph Symposium on Computer Animation (2003) Handrix: Animating the Human Hand , 2003 .

[18]  Jun Rekimoto,et al.  SmartSkin: an infrastructure for freehand manipulation on interactive surfaces , 2002, CHI.

[19]  Meredith Ringel Morris,et al.  ShadowGuides: visualizations for in-situ learning of multi-touch and whole-hand gestures , 2009, ITS '09.

[20]  Shota Yamanaka,et al.  ScraTouch , 2020, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[21]  Dominik Schmidt,et al.  HandsDown: hand-contour-based user identification for interactive surfaces , 2010, NordiCHI.

[22]  Karthik Ramani,et al.  Extended multitouch: recovering touch posture and differentiating users using a depth camera , 2012, UIST.

[23]  Xiang Cao,et al.  ShapeTouch: Leveraging contact shape on interactive surfaces , 2008, 2008 3rd IEEE International Workshop on Horizontal Interactive Human Computer Systems.

[24]  Andrew W. Fitzgibbon,et al.  Accurate, Robust, and Flexible Real-time Hand Tracking , 2015, CHI.

[25]  Mircea Nicolescu,et al.  Vision-based hand pose estimation: A review , 2007, Comput. Vis. Image Underst..

[26]  Ravin Balakrishnan,et al.  Keepin' it real: pushing the desktop metaphor with physics, piles and the pen , 2006, CHI.

[27]  Sebastian Boring,et al.  Designing user-, hand-, and handpart-aware tabletop interactions with the TouchID toolkit , 2011, ITS '11.

[28]  Abdulmotaleb El Saddik,et al.  Digital Twins: The Convergence of Multimedia Technologies , 2018, IEEE MultiMedia.

[29]  Mike Wu,et al.  Multi-finger and whole hand gestural interaction techniques for multi-user tabletop displays , 2003, UIST '03.

[30]  Saul Greenberg,et al.  Multimodal multiplayer tabletop gaming , 2007, CIE.

[31]  Vincent Lepetit,et al.  DeepPrior++: Improving Fast and Accurate 3D Hand Pose Estimation , 2017, 2017 IEEE International Conference on Computer Vision Workshops (ICCVW).

[32]  Joseph A. Paradiso,et al.  A Usability User Study Concerning Free-Hand Microgesture and Wrist-Worn Sensors , 2014, 2014 11th International Conference on Wearable and Implantable Body Sensor Networks.

[33]  Chengde Wan,et al.  MEgATrack , 2020, ACM Trans. Graph..

[34]  Jan O. Borchers,et al.  Use the Force Picker, Luke: Space-Efficient Value Input on Force-Sensitive Mobile Touchscreens , 2018, CHI.

[35]  Ravin Balakrishnan,et al.  DualKey: Miniature Screen Text Entry via Finger Identification , 2016, CHI.

[36]  Robert Xiao,et al.  Estimating 3D Finger Angle on Commodity Touchscreens , 2015, ITS.

[37]  Chris Harrison,et al.  Using shear as a supplemental two-dimensional input channel for rich touchscreen interaction , 2012, CHI.

[38]  Sven Mayer,et al.  Improving the Input Accuracy of Touchscreens using Deep Learning , 2019, CHI Extended Abstracts.

[39]  Robert Xiao,et al.  CapAuth: Identifying and Differentiating User Handprints on Commodity Capacitive Touchscreens , 2015, ITS.

[40]  Niels Henze,et al.  Investigating Unintended Inputs for One-Handed Touch Interaction Beyond the Touchscreen , 2019, MobileHCI.

[41]  Carl Gutwin,et al.  Improving Discoverability and Expert Performance in Force-Sensitive Text Selection for Touch Devices with Mode Gauges , 2018, CHI.

[42]  Ravin Balakrishnan,et al.  Pressure widgets , 2004, CHI.

[43]  Kenton O'Hara,et al.  Pre-Touch Sensing for Mobile Interaction , 2016, CHI.

[44]  Matthew S. Reynolds,et al.  Finding Common Ground: A Survey of Capacitive Sensing in Human-Computer Interaction , 2017, CHI.

[45]  Chris Harrison,et al.  TapSense: enhancing finger interaction on touch surfaces , 2011, UIST.

[46]  Gunnar Farnebäck,et al.  Two-Frame Motion Estimation Based on Polynomial Expansion , 2003, SCIA.

[47]  Niels Henze,et al.  InfiniTouch: Finger-Aware Interaction on Fully Touch Sensitive Smartphones , 2018, UIST.

[48]  Meredith Ringel Morris,et al.  User-defined gestures for surface computing , 2009, CHI.

[49]  Kristofer S. J. Pister,et al.  Acceleration sensing glove (ASG) , 1999, Digest of Papers. Third International Symposium on Wearable Computers.

[50]  Ravin Balakrishnan,et al.  Porous Interfaces for Small Screen Multitasking using Finger Identification , 2016, UIST.

[51]  Robert Xiao,et al.  TouchTools: leveraging familiarity and skill with physical tools to augment touch interaction , 2014, CHI.

[52]  Niels Henze,et al.  Investigating the feasibility of finger identification on capacitive touchscreens using deep learning , 2019, IUI.

[53]  Chen Liang,et al.  HandSee: Enabling Full Hand Interaction on Smartphone with Front Camera-based Stereo Vision , 2019, CHI.

[54]  Fabrice Matulic,et al.  Unimanual Pen+Touch Input Using Variations of Precision Grip Postures , 2018, UIST.

[55]  Jun Rekimoto,et al.  Z-touch: an infrastructure for 3d gesture interaction in the proximity of tabletop surfaces , 2010, ITS '10.

[56]  Andreas Bulling,et al.  KnuckleTouch: Enabling Knuckle Gestures on Capacitive Touchscreens using Deep Learning , 2019, Mensch & Computer.

[57]  Valentin Schwind,et al.  Touch with foreign hands: the effect of virtual hand appearance on visual-haptic integration , 2018, SAP.

[58]  Pyeong-Gook Jung,et al.  A Wearable Gesture Recognition Device for Detecting Muscular Activities Based on Air-Pressure Sensors , 2015, IEEE Transactions on Industrial Informatics.

[59]  M. Sheelagh T. Carpendale,et al.  Gestures in the wild: studying multi-touch gesture sequences on interactive tabletop exhibits , 2011, CHI.

[60]  Guy Weinzapfel,et al.  One-point touch input of vector information for computer displays , 1978, SIGGRAPH '78.

[61]  Robert Xiao,et al.  CapCam: Enabling Rapid, Ad-Hoc, Position-Tracked Interactions Between Devices , 2016, ISS.

[62]  Senaka Buthpitiya,et al.  Bodyprint: Biometric User Identification on Mobile Devices Using the Capacitive Touchscreen to Scan Body Parts , 2015, CHI.

[63]  Ian Oakley,et al.  The Flat Finger: Exploring Area Touches on Smartwatches , 2016, CHI.

[64]  Sven Mayer,et al.  Force Touch Detection on Capacitive Sensors using Deep Neural Networks , 2019, MobileHCI.

[65]  Zhensong Zhang,et al.  Left and right hand distinction for multi-touch tabletop interactions , 2014, IUI.

[66]  Itiro Siio,et al.  Ohmic-Sticker: Force-to-Motion Type Input Device that Extends Capacitive Touch Surface , 2019, UIST.

[67]  Jonna Häkkilä,et al.  Exploring finger specific touch screen interaction for mobile phone user interfaces , 2014, OZCHI.