TapSnoop: Leveraging Tap Sounds to Infer Tapstrokes on Touchscreen Devices

We propose a novel tapstroke inference attack method, called TapSnoop, that precisely recovers what user types on touchscreen devices. Inferring tapstrokes is challenging owing to 1) low tapstroke intensity and 2) dynamically-changing noise. We address these challenges by revealing the unique characteristics of tapstrokes from audio recordings exploited by TapSnoop as a side channel of tapstrokes. In particular, we develop tapstroke detection and localization algorithms that collectively leverage audio features obtained from multiple microphones, which are designed to reflect the core properties of tapstrokes. Furthermore, we improve its robustness against environmental changes, by developing environment-adaptive classification and noise subtraction algorithms. Extensive experiments with ten real-world users on both number and QWERTY keyboards show that TapSnoop can achieve an inference accuracy of 85.4% and 75.6% (96.2% and 90.8% in best case scenarios) in stable environments, respectively. TapSnoop can also achieve a reasonable accuracy even with varying noise. For example, it shows an inference accuracy of 84.8% and 72.7% in a numeric keyboard when the noise level is varied from 37.9 to 51.2 dBA and 46.7 to 60.0 dBA, respectively.

[1]  He Wang,et al.  MoLe: Motion Leaks through Smartwatch Sensors , 2015, MobiCom.

[2]  Klara Nahrstedt,et al.  Mobile Devices based Eavesdropping of Handwriting , 2020, IEEE Transactions on Mobile Computing.

[3]  Tao Gu,et al.  GazeRevealer: Inferring Password Using Smartphone Front Camera , 2018, MobiQuitous.

[4]  David Chu,et al.  SwordFight: enabling a new class of phone-to-phone action games on commodity phones , 2012, MobiSys '12.

[5]  Jeff Yan,et al.  Hearing your touch: A new acoustic side channel on smartphones , 2019, ArXiv.

[6]  Desney S. Tan,et al.  FingerIO: Using Active Sonar for Fine-Grained Finger Tracking , 2016, CHI.

[7]  Guevara Noubir,et al.  Single-stroke language-agnostic keylogging using stereo-microphones and domain specific machine learning , 2014, WiSec '14.

[8]  Lili Qiu,et al.  CAT: high-precision acoustic motion tracking , 2016, MobiCom.

[9]  David M. W. Powers,et al.  Evaluation: from precision, recall and F-measure to ROC, informedness, markedness and correlation , 2011, ArXiv.

[10]  Tao Li,et al.  EyeTell: Video-Assisted Touchscreen Keystroke Inference from Eye Movements , 2018, 2018 IEEE Symposium on Security and Privacy (SP).

[11]  Xiangyu Liu,et al.  When Good Becomes Evil: Keystroke Inference with Smartwatch , 2015, CCS.

[12]  Sangki Yun,et al.  Turning a Mobile Device into a Mouse in the Air , 2015, MobiSys.

[13]  Robert Xiao,et al.  Toffee: enabling ad hoc, around-device interaction with acoustic time-of-arrival correlation , 2014, MobileHCI '14.

[14]  Jingchang Huang,et al.  Single-Mode Wild Area Surveillance Sensor With Ultra-Low Power Design Based on Microphone Array , 2019, IEEE Access.

[15]  Jingchang Huang,et al.  A classification method for moving targets in the wild based on microphone array and linear sparse auto-encoder , 2017, Neurocomputing.

[16]  Bernhard E. Boser,et al.  A training algorithm for optimal margin classifiers , 1992, COLT '92.

[17]  Chih-Jen Lin,et al.  LIBSVM: A library for support vector machines , 2011, TIST.

[18]  Xin Sun,et al.  TextLogger: inferring longer inputs on touch screen using motion sensors , 2015, WISEC.

[19]  Duncan Hodges,et al.  Reconstructing What You Said: Text Inference Using Smartphone Motion , 2019, IEEE Transactions on Mobile Computing.

[20]  L. H. Anauer,et al.  Speech Analysis and Synthesis by Linear Prediction of the Speech Wave , 2000 .

[21]  Otmar Hilliges,et al.  In-air gestures around unmodified mobile devices , 2014, UIST.

[22]  Panlong Yang,et al.  WordRecorder: Accurate Acoustic-based Handwriting Recognition Using Deep Learning , 2018, IEEE INFOCOM 2018 - IEEE Conference on Computer Communications.

[23]  Jie Yang,et al.  Snooping Keystrokes with mm-level Audio Ranging on a Single Phone , 2015, MobiCom.

[24]  Xinyu Zhang,et al.  Ubiquitous keyboard for small mobile devices: harnessing multipath fading for fine-grained keystroke localization , 2014, MobiSys.

[25]  Tom Page,et al.  Usability of text input interfaces in smartphones , 2013 .

[26]  Zhi Xu,et al.  TapLogger: inferring user inputs on smartphone touchscreens using on-board motion sensors , 2012, WISEC '12.

[27]  Romit Roy Choudhury,et al.  Tapprints: your finger taps have fingerprints , 2012, MobiSys '12.

[28]  Marco Gruteser,et al.  VibSense: Sensing Touches on Ubiquitous Surfaces through Vibration , 2017, 2017 14th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON).

[29]  Insik Shin,et al.  UbiTap: Leveraging Acoustic Dispersion for Ubiquitous Touch Interface on Solid Surfaces , 2018, SenSys.

[30]  Hao Chen,et al.  TouchLogger: Inferring Keystrokes on Touch Screen from Smartphone Motion , 2011, HotSec.

[31]  David Chu,et al.  On the feasibility of real-time phone-to-phone 3D localization , 2011, SenSys.

[32]  John Platt,et al.  Probabilistic Outputs for Support vector Machines and Comparisons to Regularized Likelihood Methods , 1999 .

[33]  Xiaolin Li,et al.  Guoguo: enabling fine-grained indoor localization via smartphone , 2013, MobiSys '13.

[34]  Jingchang Huang,et al.  Design of an Acoustic Target Intrusion Detection System Based on Small-Aperture Microphone Array , 2017, Sensors.