Smart Touch: Improving Touch Accuracy for People with Motor Impairments with Template Matching

We present two contributions toward improving the accessibility of touch screens for people with motor impairments. First, we provide an exploration of the touch behaviors of 10 people with motor impairments, e.g., we describe how touching with the back or sides of the hand, with multiple fingers, or with knuckles creates varied multi-point touches. Second, we introduce Smart Touch, a novel template-matching technique for touch input that maps any number of arbitrary contact-areas to a user's intended (x,y) target location. The result is that users with motor impairments can touch however their abilities allow, and Smart Touch will resolve their intended touch point. Smart Touch therefore allows users to touch targets in whichever ways are most comfortable and natural for them. In an experimental evaluation, we found that Smart Touch predicted (x,y) coordinates of the users' intended target locations over three times closer to the intended target than the native Land-on and Lift-off techniques reported by the built-in touch sensors found in the Microsoft PixelSense interactive tabletop. This result is an important step toward improving touch accuracy for people with motor impairments and others for whom touch screen operation was previously impossible.

[1]  Jacob O. Wobbrock,et al.  A general-purpose target-aware pointing enhancement using pixel-level analysis of graphical interfaces , 2012, CHI.

[2]  Brad A. Myers,et al.  EdgeWrite: a stylus-based text entry method designed for high accuracy and stability of motion , 2003, UIST '03.

[3]  Douglas A. Wiegmann,et al.  The Effect of Disability and Approach on Touch Screen Performance during a Number Entry Task , 2010 .

[4]  Krzysztof Z. Gajos,et al.  Automatically generating user interfaces adapted to users' motor and vision capabilities , 2007, UIST.

[5]  Susumu Harada,et al.  The angle mouse: target-agnostic dynamic gain adjustment based on angular deviation , 2009, CHI.

[6]  Daryl Weir,et al.  Machine learning models for uncertain interaction , 2012, UIST Adjunct Proceedings '12.

[7]  Patrick Baudisch,et al.  Precise selection techniques for multi-touch screens , 2006, CHI.

[8]  Patrick Baudisch,et al.  The generalized perceived input point model and how to double touch accuracy by extracting fingerprints , 2010, CHI.

[9]  R C Littell,et al.  Statistical analysis of repeated measures data using SAS procedures. , 1998, Journal of animal science.

[10]  Shumin Zhai,et al.  High precision touch screen interaction , 2003, CHI '03.

[11]  Ben Shneiderman,et al.  High Precision Touchscreens: Design Strategies and Comparisons with a Mouse , 1991, Int. J. Man Mach. Stud..

[12]  Jacob O. Wobbrock,et al.  Enhanced area cursors: reducing fine pointing demands for people with motor impairments , 2010, UIST.

[13]  Radu-Daniel Vatavu,et al.  Gestures as point clouds: a $P recognizer for user interface prototypes , 2012, ICMI '12.

[14]  Jon Froehlich,et al.  Barrier pointing: using physical edges to assist target acquisition on mobile device touch screens , 2007, Assets '07.

[15]  Kyle Montague,et al.  Motor-impaired touchscreen interactions in the wild , 2014, ASSETS.

[16]  Joaquim A. Jorge,et al.  Assessing mobile touch interfaces for tetraplegics , 2010, Mobile HCI.

[17]  Patrick Baudisch,et al.  Lucid touch: a see-through mobile device , 2007, UIST.

[18]  Thomas G. Dietterich What is machine learning? , 2020, Archives of Disease in Childhood.

[19]  Scott E. Hudson,et al.  Monte carlo methods for managing interactive state, action and feedback under uncertainty , 2011, UIST '11.

[20]  Joaquim A. Jorge,et al.  Towards accessible touch interfaces , 2010, ASSETS '10.

[21]  Lisa Anthony,et al.  Analyzing user-generated youtube videos to understand touchscreen use by people with motor impairments , 2013, CHI.

[22]  Shumin Zhai,et al.  Bayesian touch: a statistical criterion of target selection with finger touch , 2013, UIST.

[23]  Richard E. Ladner,et al.  Freedom to roam: a study of mobile device adoption and accessibility for people with visual and motor disabilities , 2009, Assets '09.

[24]  Benjamin B. Bederson,et al.  Target size study for one-handed thumb use on small touchscreen devices , 2006, Mobile HCI.

[25]  Jacob O. Wobbrock,et al.  Improving Pointing in Graphical User Interfaces for People with Motor Impairments Through Ability-Based Design , 2013 .

[26]  Krzysztof Z. Gajos,et al.  Improving the performance of motor-impaired users with automatically-generated, ability-based interfaces , 2008, CHI.

[27]  Krzysztof Z. Gajos,et al.  Ability-Based Design: Concept, Principles and Examples , 2011, TACC.

[28]  Shumin Zhai,et al.  FFitts law: modeling finger touch with fitts' law , 2013, CHI.

[29]  Daniel Vogel,et al.  Shift: a technique for operating pen-based interfaces using touch , 2007, CHI.

[30]  Karen B. Chen,et al.  Touch screen performance by individuals with and without motor control disabilities. , 2013, Applied ergonomics.

[31]  Jacob O. Wobbrock,et al.  Taming wild behavior: the input observer for obtaining text entry and mouse pointing measures from everyday computer use , 2012, CHI.

[32]  Shari Trewin,et al.  Physical accessibility of touchscreen smartphones , 2013, ASSETS.

[33]  Ravin Balakrishnan,et al.  "Beating" Fitts' law: virtual enhancements for pointing facilitation , 2004, Int. J. Hum. Comput. Stud..

[34]  S. Holm A Simple Sequentially Rejective Multiple Test Procedure , 1979 .

[35]  Patrick Baudisch,et al.  Understanding touch , 2011, CHI.

[36]  Yang Li,et al.  Gestures without libraries, toolkits or training: a $1 recognizer for user interface prototypes , 2007, UIST.

[37]  Kyle Montague,et al.  Designing for individuals: usable touch-screen interaction through shared user models , 2012, ASSETS '12.

[38]  Scott E. Hudson,et al.  An Architecture for Generating Interactive Feedback in Probabilistic User Interfaces , 2015, CHI.

[39]  Markus Löchtefeld,et al.  A user-specific machine learning approach for improving touch accuracy on mobile devices , 2012, UIST '12.

[40]  Mary E Sesto,et al.  Performance and touch characteristics of disabled and non-disabled participants during a reciprocal tapping task using touch screen technology. , 2012, Applied ergonomics.

[41]  Christopher M. Schlick,et al.  Evaluating swabbing: a touchscreen input method for elderly users with tremor , 2011, CHI.

[42]  Leah Findlater,et al.  Accessibility in context: understanding the truly mobile experience of smartphone users with motor impairments , 2014, ASSETS.

[43]  B. Shneiderman,et al.  Improving the accuracy of touch screens: an experimental evaluation of three strategies , 1988, CHI '88.

[44]  Patrick Langdon,et al.  Developing Multimodal Adaptation Algorithm for Mobility Impaired Users by Evaluating Their Hand Strength , 2012, Int. J. Hum. Comput. Interact..

[45]  Scott E. Hudson,et al.  A framework for robust and flexible handling of inputs with uncertainty , 2010, UIST.