Overview of auditory representations in human-machine interfaces

In recent years, a large number of research projects have focused on the use of auditory representations in a broadened scope of application scenarios. Results in such projects have shown that auditory elements can effectively complement other modalities not only in the traditional desktop computer environment but also in virtual and augmented reality, mobile platforms, and other kinds of novel computing environments. The successful use of auditory representations in this growing number of application scenarios has in turn prompted researchers to rediscover the more basic auditory representations and extend them in various directions. The goal of this article is to survey both classical auditory representations (e.g., auditory icons and earcons) and those auditory representations that have been created as extensions to earlier approaches, including speech-based sounds (e.g., spearcons and spindex representations), emotionally grounded sounds (e.g., auditory emoticons and spemoticons), and various other sound types used to provide sonifications in practical scenarios. The article concludes by outlining the latest trends in auditory interface design and providing examples of these trends.

[1]  Elizabeth M. Wenzel,et al.  Localization in Virtual Acoustic Displays , 1992, Presence: Teleoperators & Virtual Environments.

[2]  Sylvain Le Groux,et al.  Interactive Sonification of the Spatial Behavior of Human and Synthetic Characters in a Mixed-Reality Environment , 2007 .

[3]  Stephen Brewster,et al.  An investigation of using music to provide navigation cues , 1998 .

[4]  Helen Petrie,et al.  The use of non-speech sounds in non-visual interfaces to the MS-Windows GUI for blind computer users , 1998 .

[5]  Marion Hersh,et al.  Perception, the Eye and Assistive Technology Issues , 2008 .

[6]  Azeddine Beghdadi,et al.  An audiodisplay tool for visually impaired people: the sound screen system , 2001, Proceedings of the Sixth International Symposium on Signal Processing and its Applications (Cat.No.01EX467).

[7]  William W. Gaver Auditory Icons: Using Sound in Computer Interfaces , 1986, Hum. Comput. Interact..

[8]  Nikolaos G. Bourbakis,et al.  Wearable Obstacle Avoidance Electronic Travel Aids for Blind: A Survey , 2010, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).

[9]  James R. Marston,et al.  Personal Guidance System for People with Visual Impairment: A Comparison of Spatial Displays for Route Guidance , 2005, Journal of visual impairment & blindness.

[10]  Stephen Brewster,et al.  Providing a Structured Method for Integrating Non-Speech Audio into Human-Computer Interfaces , 1994 .

[11]  Andy Hunt,et al.  Guest Editors' Introduction: An Introduction to Interactive Sonification , 2005, IEEE Multim..

[12]  György Wersényi EVALUATION OF AUDITORY REPRESENTATIONS FOR SELECTED APPLICATIONS OF A GRAPHICAL USER INTERFACE , 2009 .

[13]  Péter Baranyi,et al.  The Spiral Discovery Method: An Interpretable Tuning Model for CogInfoCom Channels , 2012, J. Adv. Comput. Intell. Intell. Informatics.

[14]  G. Wersenyi,et al.  Effect of Emulated Head-Tracking for Reducing Localization Errors in Virtual Audio Simulation , 2009, IEEE Transactions on Audio, Speech, and Language Processing.

[15]  Dalbir Singh Hybrid Auditory Based Interaction Framework for Driver Assistance System , 2010 .

[16]  Frank Dellaert,et al.  SWAN: System for Wearable Audio Navigation , 2007, 2007 11th IEEE International Symposium on Wearable Computers.

[17]  Tamás Gábor Csapó,et al.  Spemoticons: Text to Speech Based Emotional Auditory Cues , 2011 .

[18]  Stephen Barrass,et al.  TaDa! Demonstrations of auditory information design , 1996 .

[19]  Stephen Barrass,et al.  Auditory information design , 1998 .

[20]  Thomas Hermann,et al.  Sonification for Exploratory Data Analysis , 2002 .

[21]  M. Turvey,et al.  Hearing shape. , 2000, Journal of experimental psychology. Human perception and performance.

[22]  Dinesh K. Pai,et al.  FoleyAutomatic: physically-based sound effects for interactive simulation and animation , 2001, SIGGRAPH.

[23]  Andy Hunt,et al.  The Importance of Interaction in Sonification , 2004, ICAD.

[24]  William W. Gaver The SonicFinder: An Interface That Uses Auditory Icons , 1989, Hum. Comput. Interact..

[25]  Nina Schaffert,et al.  An investigation of online acoustic information for elite rowers in on-water training conditions , 2011 .

[26]  Thomas Hermann,et al.  Sonified Aerobics - Interactive Sonification of Coordinated Body Movements , 2011 .

[27]  György Wersényi,et al.  Localization in a HRTF-based Minimum-Audible-Angle Listening test for GUIB applications , 2007 .

[28]  Myounghoon Jeon,et al.  “Spindex”: Accelerated Initial Speech Sounds Improve Navigation Performance in Auditory Menus , 2009 .

[29]  György Wersényi Auditory Representations of a Graphical User Interface for a Better Human-Computer Interaction , 2009, CMMR/ICAD.

[30]  C. Carello,et al.  Perception of Object Length by Sound , 1998 .

[31]  Jacques Bertin,et al.  Graphics and graphic information-processing , 1981 .

[32]  Stephen Barrass,et al.  EarBenders : using stories about listening to design auditory interfaces , 1996 .

[33]  Bruce N. Walker,et al.  Navigation Performance With a Virtual Auditory Display: Effects of Beacon Sound, Capture Radius, and Practice , 2006, Hum. Factors.

[34]  Salvatore Graziani,et al.  Multisensor Strategies to Assist Blind People: A Clear-Path Indicator , 2009, IEEE Transactions on Instrumentation and Measurement.

[35]  D. Norman,et al.  Everyday listening and auditory icons , 1988 .

[36]  Jean-Pierre Gagné,et al.  The effects of blurred vision on auditory-visual speech perception in younger and older adults , 2010, International journal of audiology.

[37]  Mta Sztaki,et al.  Overview of Auditory Representations in Human-Machine Interfaces , 2013 .

[38]  Nina Schaffert,et al.  A Sound Design for Acoustic Feedback in Elite Sports , 2009, CMMR/ICAD.

[39]  Till Bovermann Tangible auditory interfaces: combining auditory displays and tangible interfaces , 2009 .

[40]  Linda Roberts,et al.  Musical vs. real world feedback signals , 1995, CHI '95.

[41]  C. Trullemans,et al.  A real-time experimental prototype for enhancement of vision rehabilitation using auditory substitution , 1998, IEEE Transactions on Biomedical Engineering.

[42]  David W. Schloerb,et al.  BlindAid: A learning environment for enabling people who are blind to explore and navigate through unknown real spaces , 2008, 2008 Virtual Rehabilitation.

[43]  Nasser Sherkat,et al.  Enabling the use of windows environment by the blind and partially sighted , 1993 .

[44]  Pawel Strumillo,et al.  Measurement System for Personalized Head-Related Transfer Functions and Its Verification by Virtual Source Localization Trials with Visually Impaired and Sighted Individuals , 2010 .

[45]  M. Goldberg,et al.  Formalizing cognitive and motor strategy of haptic exploratory movements of individuals who are blind , 2004, Proceedings. Second International Conference on Creating, Connecting and Collaborating through Computing.

[46]  S.A.K. Aljunid,et al.  An electronically guided walking stick for the blind , 2001, 2001 Conference Proceedings of the 23rd Annual International Conference of the IEEE Engineering in Medicine and Biology Society.

[47]  Bruce N. Walker,et al.  Listener, Task, and Auditory Graph: Toward a Conceptual Model of Auditory Graph Comprehension , 2007 .

[48]  Davide Rocchesso,et al.  Sounding Objects , 2003, IEEE Multim..

[49]  Gregory L. Goodrich,et al.  A Follow-up Study of the Mowat Sensor's Applications, Frequency of Use, and Maintenance Reliability , 1981 .

[50]  B. Ando,et al.  A Smart Multisensor Approach to Assist Blind People in Specific Urban Navigation Tasks , 2008, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[51]  Stephen A. Brewster,et al.  Parallel earcons: reducing the length of audio messages , 1995, Int. J. Hum. Comput. Stud..

[52]  Manne-Sakari Mustonen,et al.  A review-based conceptual analysis of auditory signs and their design , 2008 .

[53]  Stephen A. Brewster,et al.  Using nonspeech sounds to provide navigation cues , 1998, TCHI.

[54]  José González,et al.  Multichannel audio biofeedback for dynamical coupling between prosthetic hands and their users , 2010, Ind. Robot.

[55]  György Wersényi,et al.  Virtual Localization by Blind Persons , 2012 .

[56]  Henrik Møller Fundamentals of binaural technology , 1991 .

[57]  Meera Blattner,et al.  Earcons and Icons: Their Structure and Common Design Principles , 1989, Hum. Comput. Interact..

[58]  Davide Rocchesso,et al.  The Sonification Handbook , 2011 .

[59]  Michael J. Massimino,et al.  Improved force perception through sensory substitution , 1995 .

[60]  Perry R. Cook,et al.  An auditory display system for aiding interjoint coordination , 2000 .

[61]  Myounghoon Jeon,et al.  Spindex (Speech Index) Improves Auditory Menu Acceptance and Navigation Performance , 2011, TACC.

[62]  R. R. Patterson,et al.  Guidelines for auditory warning systems on civil aircraft , 1982 .

[63]  Ab de Haan,et al.  When it Sounds like a Duck and it Looks like a Dog… Auditory icons vs. Earcons in Multimedia Environments , 2000 .

[64]  Homayoun Najjaran,et al.  The EMPATHY MACHINE , 2012, 2012 IEEE International Conference on Systems, Man, and Cybernetics (SMC).

[65]  J Edworthy,et al.  Improving Auditory Warning Design: Quantifying and Predicting the Effects of Different Warning Parameters on Perceived Urgency , 1993, Human factors.

[66]  Helen Petrie,et al.  Providing Acces to Graphical-Based User Interfaces (GUIs) for Blind People Using a Multimedia System Based on Spatial Audio Representation , 1993 .

[67]  Maria Ebling Virtual Senses , 2009, IEEE Pervasive Computing.

[68]  Michael Friis Sørensen,et al.  Head-Related Transfer Functions of Human Subjects , 1995 .

[69]  William W. Gaver,et al.  Sound Support for Collaboration , 1991, ECSCW.

[70]  Judy Edworthy,et al.  Learning auditory warnings: The effects of sound type, verbal labelling and imagery on the identification of alarm sounds , 1999 .

[71]  Jung-Kyong Kim,et al.  Generalized Learning of Visual-to-auditory Substitution in Sighted Individuals , 2008 .

[72]  Brian F. G. Katz,et al.  Morphocons: A New Sonification Concept Based on Morphological Earcons , 2012 .

[73]  Peter Baranyi,et al.  Perceptual Interpolation and Open-Ended Exploration of Auditory Icons and Earcons , 2011 .

[74]  F L Wightman,et al.  Localization using nonindividualized head-related transfer functions. , 1993, The Journal of the Acoustical Society of America.

[75]  共立出版株式会社 コンピュータ・サイエンス : ACM computing surveys , 1978 .

[76]  Nina Schaffert,et al.  Modeling the Rowing Stroke Cycle Acoustically , 2012 .

[77]  Maximo Cobos,et al.  Elevation in Wave-Field Synthesis Using HRTF Cues , 2010 .

[78]  G. Meffe Techno‐Arrogance and Halfway Technologies: Salmon Hatcheries on the Pacific Coast of North America , 1992 .

[79]  Bruce N. Walker,et al.  SPEARCONS: SPEECH-BASED EARCONS IMPROVE NAVIGATION PERFORMANCE IN AUDITORY MENUS , 2006 .

[80]  Christopher Frauenberger,et al.  Auditory display design - An investigation of a design pattern approach , 2009, Int. J. Hum. Comput. Stud..

[81]  Raymond M. Fish An Audio Display for the Blind , 1976, IEEE Transactions on Biomedical Engineering.

[82]  J Edworthy,et al.  Improving Auditory Warning Design: Relationship between Warning Sound Parameters and Perceived Urgency , 1991, Human factors.

[83]  Helge Ritter,et al.  Listen to your Data: Model-Based Sonification for Data Analysis , 1999 .

[84]  Catherine Guastavino,et al.  USABILITY OF NON-SPEECH SOUNDS IN USER INTERFACES , 2008 .

[85]  Håkan Alm,et al.  Auditory signs to support traffic awareness , 2009 .

[86]  F.B. Horak,et al.  A portable audio-biofeedback system to improve postural control , 2004, The 26th Annual International Conference of the IEEE Engineering in Medicine and Biology Society.

[87]  Davide Rocchesso,et al.  Sound authoring tools for future multimedia systems , 1999, Proceedings IEEE International Conference on Multimedia Computing and Systems.

[88]  Brian Gygi,et al.  Spectral-temporal factors in the identification of environmental sounds. , 2004, The Journal of the Acoustical Society of America.

[89]  R Graham,et al.  Use of auditory icons as emergency warnings: evaluation within a vehicle collision avoidance application. , 1999, Ergonomics.

[90]  Eoin Brazil,et al.  Human-Computer Interaction Design based on Interactive Sonification - Hearing Actions or Instruments/Agents. , 2004 .

[91]  Mark R. Anderson,et al.  Direct comparison of the impact of head tracking, reverberation, and individualized head-related transfer functions on the spatial perception of a virtual speech source. , 2001, Journal of the Audio Engineering Society. Audio Engineering Society.

[92]  Cumhur Erkut,et al.  Auditory feedback in an interactive rhythmic tutoring system , 2011, AM '11.

[93]  Roberto Manduchi,et al.  Watch your head: A wearable collision warning system for the blind , 2010, 2010 IEEE Sensors.

[94]  R. Patterson Auditory warning sounds in the work environment. , 1990, Philosophical transactions of the Royal Society of London. Series B, Biological sciences.

[95]  Daniel Thalmann,et al.  A wearable system for mobility improvement of visually impaired people , 2007, The Visual Computer.

[96]  Stephen A. Brewster,et al.  Understanding concurrent earcons: Applying auditory scene analysis principles to concurrent earcon recognition , 2004, TAP.

[97]  Peter Froehlich,et al.  Expressive Text-to-Speech: A user-centred approach to sound design in voice-enabled mobile applications , 2004 .

[98]  L. Kay,et al.  Electronic Aids for Blind Persons : an Interdisciplinary Subject , 1984 .

[99]  N. Bourbakis,et al.  Sensing Surrounding 3-D Space for Navigation of the Blind , 2008, IEEE Engineering in Medicine and Biology Magazine.

[100]  Derek Brock,et al.  ENCODING URGENCY IN LEGACY AUDIO ALERTING SYSTEMS , 2005 .

[101]  György Wersényi,et al.  Localization in a HRTF-based Minimum Audible Angle Listening Test on a 2D Sound Screen for GUIB Applications , 2003 .

[102]  Luiz Claudio Locatelli Ventura,et al.  Remote guide for guiding the visually impaired , 2011, ISSNIP Biosignals and Biorobotics Conference 2011.

[103]  Sandra Pauletto,et al.  The design of an audio film for the visually impaired , 2009 .

[104]  Stephen Brewster,et al.  The sound of musicons: investigating the design of musically derived audio cues , 2012 .

[105]  György Wersényi EVALUATION OF USER HABITS FOR CREATING AUDITORY REPRESENTATIONS OF DIFFERENT SOFTWARE APPLICATIONS FOR BLIND PERSONS , 2008 .

[106]  György Wersényi,et al.  Localization in a Head-Related Transfer Function-based virtual audio synthesis using additional high-pass and low-pass filtering of sound sources , 2007 .

[107]  Elizabeth D. Mynatt Transforming Graphical Interfaces Into Auditory Interfaces for Blind Users , 1997, Hum. Comput. Interact..

[108]  Brian Gygi Studying environmental sounds the Watson way , 2004 .

[109]  Helge Ritter,et al.  TANGIBLE DATA SCANNING SONIFICATION MODEL , 2006 .

[110]  Stephen A. Brewster,et al.  Reminders that make sense: Designing multimodal notifications for the home , 2011, 2011 5th International Conference on Pervasive Computing Technologies for Healthcare (PervasiveHealth) and Workshops.

[111]  Thomas B. Sheridan,et al.  Sensory Substitution for Force Feedback in Teleoperation , 1992, Presence: Teleoperators & Virtual Environments.

[112]  Laurie M. Heller,et al.  When sound effects are better than the real thing , 2002 .

[113]  F. McKiel Audio-enabled graphical user interface for the blind or visually impaired , 1992, Proceedings of the Johns Hopkins National Search for Computing Applications to Assist Persons with Disabilities.

[114]  Tilman Dingler,et al.  Learnabiltiy of Sound Cues for Environmental Features: Auditory Icons, Earcons, Spearcons, and Speech , 2008 .

[115]  N. A. Bradley,et al.  Assistive Technology For Visually Impaired And Blind People , 2008 .

[116]  G. Vanderheiden,et al.  The Graphical User Interface: Crisis, Danger, and Opportunity , 1990 .

[117]  Maria Klara Wolters,et al.  Name that tune: musicons as reminders in the home , 2011, CHI.

[118]  Jeffrey E. Boyd,et al.  CORRECTIVE SONIC FEEDBACK FOR SPEED SKATING: A CASE STUDY , 2010 .

[119]  J. Ballas Common factors in the identification of an assortment of brief everyday sounds. , 1993, Journal of experimental psychology. Human perception and performance.

[120]  Brian Gygi,et al.  From Signal to Substance and Back: Insights from Environmental Sound Research to Auditory Display Design , 2009, CMMR/ICAD.

[121]  Peter B. L. Meijer,et al.  An experimental system for auditory image representations , 1992, IEEE Transactions on Biomedical Engineering.

[122]  William W. Gaver Synthesizing auditory icons , 1993, INTERCHI.

[123]  Gregory H. Wakefield,et al.  Introduction to Head-Related Transfer Functions (HRTFs): Representations of HRTFs in Time, Frequency, and Space , 2001 .

[124]  D. Basta,et al.  Vestibular rehabilitation by auditory feedback in otolith disorders. , 2008, Gait & posture.

[125]  Bruce N. Walker,et al.  LEARNING RATES FOR AUDITORY MENUS ENHANCED WITH SPEARCONS VERSUS EARCONS , 2007 .

[126]  Michael A. Nees,et al.  Memory for Auditory Icons and Earcons with Localization Cues , 2007 .

[127]  Chen Shen,et al.  Synthesizing sounds from rigid-body simulations , 2002, SCA '02.

[128]  R. M. Schafer,et al.  The tuning of the world , 1977 .

[129]  Gregory Kramer,et al.  Auditory Display: Sonification, Audification, And Auditory Interfaces , 1994 .