Multiple spatial sounds in hierarchical menu navigation for visually impaired computer users

This paper describes a user study on the benefits and drawbacks of simultaneous spatial sounds in auditory interfaces for visually impaired and blind computer users. Two different auditory interfaces in spatial and non-spatial condition were proposed to represent the hierarchical menu structure of a simple word processing application. In the horizontal interface, the sound sources or the menu items were located in the horizontal plane on a virtual ring surrounding the user's head, while the sound sources in the vertical interface were aligned one above the other in front of the user. In the vertical interface, the central pitch of the sound sources at different elevations was changed in order to improve the otherwise relatively low localization performance in the vertical dimension. The interaction with the interfaces was based on a standard computer keyboard for input and a pair of studio headphones for output. Twelve blind or visually impaired test subjects were asked to perform ten different word processing tasks within four experiment conditions. Task completion times, navigation performance, overall satisfaction and cognitive workload were evaluated. The initial hypothesis, i.e. that the spatial auditory interfaces with multiple simultaneous sounds should prove to be faster and more efficient than non-spatial ones, was not confirmed. On the contrary-spatial auditory interfaces proved to be significantly slower due to the high cognitive workload and temporal demand. The majority of users did in fact finish tasks with less navigation and key pressing; however, they required much more time. They reported the spatial auditory interfaces to be hard to use for a longer period of time due to the high temporal and mental demand, especially with regards to the comprehension of multiple simultaneous sounds. The comparison between the horizontal and vertical interface showed no significant differences between the two. It is important to point out that all participants were novice users of the system; therefore it is possible that the overall performance could change with a more extensive use of the interfaces and an increased number of trials or experiments sets. Our interviews with visually impaired and blind computer users showed that they are used to sharing their auditory channel in order to perform multiple simultaneous tasks such as listening to the radio, talking to somebody, using the computer, etc. As the perception of multiple simultaneous sounds requires the entire capacity of the auditory channel and total concentration of the listener, it does therefore not enable such multitasking.

[1]  Jaka Sodnik,et al.  Spatial Auditory Interface for Word Processing Application , 2009, 2009 Second International Conferences on Advances in Computer-Human Interactions.

[2]  Harold R. Booher,et al.  Manprint: An Approach to Systems Integration , 1990 .

[3]  Christopher Frauenberger,et al.  Patterns in auditory menu design , 2006 .

[4]  Rahul Sawhney,et al.  Sonic Grid: an auditory interface for the visually impaired to navigate GUI-based environments , 2008, IUI '08.

[5]  Rudolf Susnik,et al.  Spatial sound resolution of an interpolated HRIR library , 2005 .

[6]  Christopher Frauenberger,et al.  SPATIAL AUDITORY DISPLAYS A STUDY ON THE USE OF VIRTUAL AUDIO ENVIRONMENTS AS INTERFACES FOR USERS WITH VISUAL DISABILITIES , 2004 .

[7]  H S Colburn,et al.  Speech intelligibility and localization in a multi-source environment. , 1999, The Journal of the Acoustical Society of America.

[8]  Simon Carlile,et al.  Contrasting monaural and interaural spectral cues for human sound localization. , 2004, The Journal of the Acoustical Society of America.

[9]  Barry Arons,et al.  A Review of The Cocktail Party Effect , 1992 .

[10]  Stephen A. Brewster,et al.  Multimodal 'eyes-free' interaction techniques for wearable devices , 2003, CHI '03.

[11]  Christopher Frauenberger,et al.  Interaction patterns for auditory user interfaces , 2005 .

[12]  Bruce N. Walker,et al.  SPEARCONS: SPEECH-BASED EARCONS IMPROVE NAVIGATION PERFORMANCE IN AUDITORY MENUS , 2006 .

[13]  Michael Cohen,et al.  Multidimensional Audio Window Management , 1991, Int. J. Man Mach. Stud..

[14]  Rudolf Susnik,et al.  An elevation coding method for auditory displays , 2008 .

[15]  Chris Schmandt,et al.  Dynamic Soundscape: mapping time to space for audio browsing , 1997, CHI.

[16]  Lisa Stifelman The cocktail party e ect in auditory interfaces: A study of simultaneous presentation , 1994 .

[17]  Bill Gardner,et al.  HRTF Measurements of a KEMAR Dummy-Head Microphone , 1994 .

[18]  Constantine Stephanidis,et al.  A generic direct-manipulation 3D-auditory environment for hierarchical navigation in non-visual interaction , 1996, Assets '96.

[19]  Christopher D. Wickens,et al.  Workload Assessment and Prediction , 1990 .

[20]  William W. Gaver Auditory Icons: Using Sound in Computer Interfaces , 1986, Hum. Comput. Interact..

[21]  C. Avendano,et al.  The CIPIC HRTF database , 2001, Proceedings of the 2001 IEEE Workshop on the Applications of Signal Processing to Audio and Acoustics (Cat. No.01TH8575).

[22]  William A. Yost,et al.  Spatial hearing: The psychophysics of human sound localization, revised edition , 1998 .

[23]  A. Bronkhorst,et al.  Multichannel speech intelligibility and talker recognition using monaural, binaural, and three-dimensional auditory presentation. , 2000, The Journal of the Acoustical Society of America.

[24]  Bruce N. Walker,et al.  Navigation Efficiency of Two Dimensional Auditory Menus Using Spearcon Enhancements , 2008 .

[25]  Chris Schmandt,et al.  AudioStreamer: exploiting simultaneity for listening , 1995, CHI 95 Conference Companion.

[26]  J. Blauert Spatial Hearing: The Psychophysics of Human Sound Localization , 1983 .

[27]  Mark Billinghurst,et al.  A user study of auditory versus visual interfaces for use while driving , 2008, Int. J. Hum. Comput. Stud..

[28]  Stephen A. Brewster,et al.  An evaluation of earcons for use in auditory human-computer interfaces , 1993, INTERCHI.

[29]  Constantine Stephanidis,et al.  A 3D-auditory environment for hierarchical navigation in non-visual interaction , 1996 .

[30]  Durand R. Begault,et al.  3-D Sound for Virtual Reality and Multimedia Cambridge , 1994 .