EarRumble: Discreet Hands- and Eyes-Free Input by Voluntary Tensor Tympani Muscle Contraction

We explore how discreet input can be provided using the tensor tympani - a small muscle in the middle ear that some people can voluntarily contract to induce a dull rumbling sound. We investigate the prevalence and ability to control the muscle through an online questionnaire (N=192) in which 43.2% of respondents reported the ability to “ear rumble”. Data collected from participants (N=16) shows how in-ear barometry can be used to detect voluntary tensor tympani contraction in the sealed ear canal. This data was used to train a classifier based on three simple ear rumble “gestures” which achieved 95% accuracy. Finally, we evaluate the use of ear rumbling for interaction, grounded in three manual, dual-task application scenarios (N=8). This highlights the applicability of EarRumble as a low-effort and discreet eyes- and hands-free interaction technique that users found “magical” and “almost telepathic”.

[1]  Austin Henderson,et al.  Making sense of sensing systems: five questions for designers and researchers , 2002, CHI.

[2]  Ravi Vaidyanathan,et al.  Tongue-Movement Communication and Control Concept for Hands-Free Human–Machine Interfaces , 2007, IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans.

[3]  S. Hart,et al.  Development of NASA-TLX (Task Load Index): Results of Empirical and Theoretical Research , 1988 .

[4]  D. Cochlear,et al.  The Acoustic Middle Ear Muscle Reflex , 1974 .

[5]  Yuanchun Shi,et al.  EarBuddy: Enabling On-Face Interaction via Wireless Earbuds , 2020, CHI.

[6]  Wen-Huang Cheng,et al.  FingerPad: private and subtle interaction using fingertips , 2013, UIST.

[7]  A. Turner A Text-Book of the Diseases of the Ear for Students and Practitioners , 1903, Edinburgh Medical Journal.

[8]  Alexander De Luca,et al.  Evaluation of eye-gaze interaction methods for security enhanced PIN-entry , 2007, OZCHI '07.

[9]  Olivier Chapuis,et al.  Using rhythmic patterns as an input method , 2012, CHI.

[10]  Carlos Tejada,et al.  Bitey: an exploration of tooth click gestures for hands-free user interface control , 2016, MobileHCI.

[11]  Pattie Maes,et al.  Byte.it: Discreet Teeth Gestures for Mobile Device Interaction , 2019, CHI Extended Abstracts.

[12]  Allen Newell,et al.  The keystroke-level model for user performance time with interactive systems , 1980, CACM.

[13]  R. Mazlan,et al.  Ear infection and hearing loss amongst headphone users. , 2002, The Malaysian journal of medical sciences : MJMS.

[14]  A. Starr,et al.  ELECTROMYOGRAPHY OF MIDDLE EAR MUSCLES IN MAN DURING MOTOR ACTIVITIES , 1963, Acta neurologica Scandinavica.

[15]  Matt Anderson,et al.  FreeDigiter: a contact-free device for gesture control , 2004, Eighth International Symposium on Wearable Computers.

[16]  Vincent P. Buil,et al.  Headphones with touch control , 2005, Mobile HCI.

[17]  Xing-Dong Yang,et al.  EarTouch: Facilitating Smartphone Use for Visually Impaired People in Mobile and Public Scenarios , 2019, CHI.

[18]  Joseph A. Paradiso,et al.  WristFlex: low-power gesture input with wrist-worn pressure sensors , 2014, UIST.

[19]  Stephen A. Brewster,et al.  Gestures all around us: user differences in social acceptability perceptions of gesture based interfaces , 2009, Mobile HCI.

[20]  Andreas W. Kempa-Liehr,et al.  Time Series FeatuRe Extraction on basis of Scalable Hypothesis tests (tsfresh - A Python package) , 2018, Neurocomputing.

[21]  Maysam Ghovanloo,et al.  The tongue and ear interface: a wearable system for silent speech recognition , 2014, SEMWEB.

[22]  Yuta Sugiura,et al.  EarTouch: turning the ear into an input surface , 2017, MobileHCI.

[23]  Kasper Hornbæk,et al.  Charting Subtle Interaction in the HCI Literature , 2019, CHI.

[24]  Kai Kunze,et al.  Itchy nose: discreet gesture interaction using EOG sensors in smart eyewear , 2017, SEMWEB.

[25]  Pourang Irani,et al.  Consumed endurance: a metric to quantify arm fatigue of mid-air interactions , 2014, CHI.

[26]  Henry Been-Lirn Duh,et al.  Understanding the effects of discreet real-time social interaction on student engagement in lectures , 2013, OZCHI.

[27]  Masaaki Fukumoto,et al.  Tap control for headphones without sensors , 2011, UIST.

[28]  Denys J. C. Matthies InEar BioFeedController: a headset for hands-free and eyes-free interaction with mobile devices , 2013, CHI Extended Abstracts.

[29]  G. Djupesland MIDDLE EAR MUSCLE REFLEXES ELICITED BY ACOUSTIC AND NONACOUSTIC STIMULATION. , 1964, Acta oto-laryngologica. Supplementum.

[30]  S. Ingelstedt,et al.  Ear drum movements following stimulation of the middle ear muscles. , 1963, Acta oto-laryngologica. Supplementum.

[31]  Enrico Costanza,et al.  Toward subtle intimate interfaces for mobile devices using an EMG controller , 2005, CHI.

[32]  Florian Alt,et al.  The Role of Eye Gaze in Security and Privacy Applications: Survey and Future HCI Research Directions , 2020, CHI.

[33]  Hitoshi Imaoka,et al.  Fast and accurate personal authentication using ear acoustics , 2016, 2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA).

[34]  Michael Rohs,et al.  A Taxonomy of Microinteractions: Defining Microgestures Based on Ergonomic and Scenario-Dependent Requirements , 2011, INTERACT.

[35]  Bodo Urban,et al.  EarFieldSensing: A Novel In-Ear Electric Field Sensing to Enrich Wearable Gesture Input through Facial Expressions , 2017, CHI.

[36]  Paul Lukowicz,et al.  Analysis of Chewing Sounds for Dietary Monitoring , 2005, UbiComp.

[37]  Junbo Wang,et al.  Magic Ring: a self-contained gesture input device on finger , 2013, MUM.

[38]  Hiroki Watanabe,et al.  Facial expression recognition using ear canal transfer function , 2019, UbiComp.

[39]  George Tzanetakis,et al.  A comparison between audio and IMU data to detect chewing events based on an earable device , 2020, AH.

[40]  Buntarou Shizuki,et al.  CanalSense: Face-Related Movement Recognition System based on Sensing Air Pressure in Ear Canals , 2017, UIST.

[41]  William A. Ahroon,et al.  Human middle-ear muscles rarely contract in anticipation of acoustic impulses: Implications for hearing risk assessments , 2019, Hearing Research.

[42]  H H Koester,et al.  Adaptive one-switch row-column scanning. , 1999, IEEE transactions on rehabilitation engineering : a publication of the IEEE Engineering in Medicine and Biology Society.

[43]  Daniel Vogel,et al.  Gunslinger: Subtle Arms-down Mid-air Interaction , 2015, UIST.

[44]  Aage R. Møller,et al.  Network Model of the Middle Ear , 1961 .

[45]  Stephen A. Brewster,et al.  Rhythmic micro-gestures: discreet interaction on-the-go , 2017, ICMI.

[46]  Thad Starner,et al.  Stick it in your ear: building an in-ear jaw movement sensor , 2015, UbiComp/ISWC Adjunct.

[47]  Suranga Nanayakkara,et al.  ChewIt. An Intraoral Interface for Discreet Interactions , 2019, CHI.

[48]  Edward D McCoul,et al.  Validating the clinical assessment of eustachian tube dysfunction: The eustachian tube dysfunction questionnaire (ETDQ‐7) , 2012, The Laryngoscope.

[49]  Max Mühlhäuser,et al.  EarPut: augmenting behind-the-ear devices for ear-based interaction , 2013, CHI Extended Abstracts.