Lip Reading-Based User Authentication Through Acoustic Sensing on Smartphones

To prevent users’ privacy from leakage, more and more mobile devices employ biometric-based authentication approaches, such as fingerprint, face recognition, voiceprint authentications, and so on, to enhance the privacy protection. However, these approaches are vulnerable to replay attacks. Although the state-of-art solutions utilize liveness verification to combat the attacks, existing approaches are sensitive to ambient environments, such as ambient lights and surrounding audible noises. Toward this end, we explore liveness verification of user authentication leveraging users’ mouth movements, which are robust to noisy environments. In this paper, we propose a lip reading-based user authentication system, LipPass, which extracts unique behavioral characteristics of users’ speaking mouths through acoustic sensing on smartphones for user authentication. We first investigate Doppler profiles of acoustic signals caused by users’ speaking mouths and find that there are unique mouth movement patterns for different individuals. To characterize the mouth movements, we propose a deep learning-based method to extract efficient features from Doppler profiles and employ softmax function, support vector machine, and support vector domain description to construct multi-class identifier, binary classifiers, and spoofer detectors for mouth state identification, user identification, and spoofer detection, respectively. Afterward, we develop a balanced binary tree-based authentication approach to accurately identify each individual leveraging these binary classifiers and spoofer detectors with respect to registered users. Through extensive experiments involving 48 volunteers in four real environments, LipPass can achieve 90.2% accuracy in user identification and 93.1% accuracy in spoofer detection.

[1]  Sangki Yun,et al.  Turning a Mobile Device into a Mouse in the Air , 2015, MobiSys.

[2]  Kaishun Wu,et al.  We Can Hear You with Wi-Fi! , 2016, IEEE Trans. Mob. Comput..

[3]  Cam-Tu Nguyen,et al.  SilentTalk: Lip reading through ultrasonic sensing on mobile phones , 2017, IEEE INFOCOM 2017 - IEEE Conference on Computer Communications.

[4]  Yoshua Bengio,et al.  Extracting and composing robust features with denoising autoencoders , 2008, ICML '08.

[5]  Nalini K. Ratha,et al.  Automatic Fingerprint Recognition Systems , 2011, Springer New York.

[6]  C. Sidney Burrus,et al.  Generalized digital Butterworth filter design , 1998, IEEE Trans. Signal Process..

[7]  Kaishun Wu,et al.  We Can Hear You with Wi-Fi! , 2014, IEEE Transactions on Mobile Computing.

[8]  Wei Wang,et al.  Device-free gesture tracking using acoustic signals , 2016, MobiCom.

[9]  L. G. Kersta Voiceprint Identification , 1962, Nature.

[10]  Kuldip K. Paliwal,et al.  Feature extraction and dimensionality reduction algorithms and their applications in vowel recognition , 2003, Pattern Recognit..

[11]  Desney S. Tan,et al.  FingerIO: Using Active Sonar for Fine-Grained Finger Tracking , 2016, CHI.

[12]  Anastasios Tefas,et al.  Using Support Vector Machines to Enhance the Performance of Elastic Graph Matching for Frontal Face Authentication , 2001, IEEE Trans. Pattern Anal. Mach. Intell..

[13]  Sangki Yun,et al.  Strata: Fine-Grained Acoustic-based Device-Free Tracking , 2017, MobiSys.

[14]  Yu Shi,et al.  SilentKey: A New Authentication Framework through Ultrasonic-based Lip Reading , 2018, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[15]  J. Yan,et al.  Password memorability and security: empirical results , 2004, IEEE Security & Privacy Magazine.

[16]  Juergen Luettin,et al.  Speaker identification by lipreading , 1996, Proceeding of Fourth International Conference on Spoken Language Processing. ICSLP '96.

[17]  M. Angela Sasse,et al.  The true cost of unusable password policies: password use in the wild , 2010, CHI.

[18]  Robert P. W. Duin,et al.  Support vector domain description , 1999, Pattern Recognit. Lett..

[19]  Bing Zhou,et al.  BatMapper: Acoustic Sensing Based Indoor Floor Plan Construction Using Smartphones , 2017, MobiSys.

[20]  G. Chetty,et al.  Multi-Level Liveness Verification for Face-Voice Biometric Authentication , 2006, 2006 Biometrics Symposium: Special Session on Research at the Biometric Consortium Conference.

[21]  Jie Yang,et al.  Hearing Your Voice is Not Enough: An Articulatory Gesture Based Liveness Detection for Voice Authentication , 2017, CCS.

[22]  Desney S. Tan,et al.  SoundWave: using the doppler effect to sense gestures , 2012, CHI.

[23]  William Stafford Noble,et al.  Support vector machine , 2013 .

[24]  Jie Yang,et al.  VoiceLive: A Phoneme Localization based Liveness Detection for Voice Authentication on Smartphones , 2016, CCS.

[25]  Eugene H. Spafford,et al.  OPUS: Preventing weak password choices , 1992, Comput. Secur..

[26]  T. Kailath,et al.  Numerically stable fast transversal filters for recursive least squares adaptive filtering , 1991, IEEE Trans. Signal Process..

[27]  B. Ripley,et al.  Pattern Recognition , 1968, Nature.

[28]  Corinna Cortes,et al.  Support-Vector Networks , 1995, Machine Learning.

[29]  Pascal Vincent,et al.  Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion , 2010, J. Mach. Learn. Res..

[30]  Paul L. Rosin,et al.  Assessing the Uniqueness and Permanence of Facial Actions for Use in Biometric Applications , 2010, IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans.

[31]  B. Kröger,et al.  A gesture‐based dynamic model describing articulatory movement data , 1995 .

[32]  Shan Suthaharan,et al.  Support Vector Machine , 2016 .