Use of binaural and monaural cues to identify the lateral position of a virtual object using echoes

Under certain conditions, sighted and blind humans can use echoes to discern characteristics of otherwise silent objects. Previous research concluded that robust horizontal-plane object localisation ability, without using head movement, depends on information above 2 kHz. While a strong interaural level difference (ILD) cue is available, it was not clear if listeners were using that or the monaural level cue that necessarily accompanies ILD. In this experiment, 13 sighted and normal-hearing listeners were asked to identify the right-vs.-left position of an object in virtual auditory space. Sounds were manipulated to remove binaural cues (binaural vs. diotic presentation) and prevent the use of monaural level cues (using level roving). With low- (<2 kHz) and high- (>2 kHz) frequency bands of noise, performance with binaural presentation and level rove exceeded that expected from use of monaural level cues and that with diotic presentation. It is argued that a high-frequency binaural cue (most likely ILD), and not a monaural level cue, is crucial for robust object localisation without head movement.

[1]  Robert S. Wall,et al.  Low Frequency Sound as a Navigational Tool for People with Visual Impairments , 2002 .

[2]  M. Lassonde,et al.  Blind subjects process auditory spectral cues more efficiently than sighted individuals , 2004, Experimental Brain Research.

[3]  S. Forbes,et al.  Frequency Discrimination as a Function of Frequency , 1967 .

[4]  Gerald Kidd,et al.  Limiting unwanted cues via random rove applied to the yes-no and multiple-alternative forced choice paradigms. , 2009, The Journal of the Acoustical Society of America.

[5]  B. Moore Cochlear hearing loss : physiological, psychological and technical issues , 2014 .

[6]  H S Colburn,et al.  Interaural correlation discrimination: i. bandwidth and level dependence. , 1981, The Journal of the Acoustical Society of America.

[7]  J. C. Middlebrooks,et al.  Listener weighting of cues for lateral angle: the duplex theory of sound localization revisited. , 2002, The Journal of the Acoustical Society of America.

[8]  M. Lutman,et al.  Learning to discriminate interaural time differences at low and high frequencies , 2007, International journal of audiology.

[9]  W. Yost Pitch and pitch discrimination of broadband signals with rippled power spectra. , 1978, The Journal of the Acoustical Society of America.

[10]  Victor Candas,et al.  Enhanced sensitivity to echo cues in blind subjects , 2005, Experimental Brain Research.

[11]  K. Arehart Cochlear Hearing Loss: Physiological, Psychological and Technical Issues, Second Edition (Wiley Series in Human Communication Science) , 2008 .

[12]  Daniel Rowan,et al.  Identification of the lateral position of a virtual object based on echoes by humans , 2013, Hearing Research.

[13]  Daniel Rowan,et al.  Identification of auditory cues utilized in human echolocation - objective measurement results , 2009, 2009 9th International Conference on Information Technology and Applications in Biomedicine.

[14]  William M Hartmann,et al.  Interaural level differences and the level-meter model. , 2002, The Journal of the Acoustical Society of America.

[15]  C. Trahiotis,et al.  Lateralization of bands of noise and sinusoidally amplitude-modulated tones: effects of spectral locus and bandwidth. , 1986, The Journal of the Acoustical Society of America.

[16]  W. Hartmann,et al.  Human interaural time difference thresholds for sine tones: the high-frequency limit. , 2013, The Journal of the Acoustical Society of America.

[17]  David Whitney,et al.  Ultrafine spatial acuity of blind expert human echolocators , 2012, Experimental Brain Research.

[18]  Brian Gygi,et al.  Individual differences in auditory abilities. , 2007, The Journal of the Acoustical Society of America.

[19]  H. Levitt,et al.  Effect of Dual Sensory Loss on Auditory Localization: Implications for Intervention , 2007, Trends in amplification.

[20]  Lore Thaler,et al.  Echolocation may have real-life advantages for blind people: an analysis of survey data , 2013, Front. Physiol..

[21]  A. Dufour,et al.  Auditory compensation in myopic humans: involvement of binaural, monaural, or echo cues? , 2005, Brain Research.

[22]  John G. Holden,et al.  A fractal approach to dynamic inference and distribution analysis , 2013, Front. Physio..

[23]  Brian C. J. Moore,et al.  A summary of research investigating echolocation abilities of blind and sighted humans , 2014, Hearing Research.

[24]  M. Goodale,et al.  Citation for Published Item: Use Policy Neural Correlates of Natural Human Echolocation in Early and Late Blind Echolocation Experts , 2022 .

[25]  F. Wightman,et al.  The dominant role of low-frequency interaural time differences in sound localization. , 1992, The Journal of the Acoustical Society of America.

[26]  Santani Teng,et al.  The acuity of echolocation: Spatial resolution in the sighted compared to expert performance. , 2011, Journal of visual impairment & blindness.

[27]  J. B. Pittenger,et al.  Human Echolocation as a Basic Form of Perception and Action , 1995 .

[28]  D. M. Green,et al.  Frequency discrimination as a function of frequency and sensation level. , 1977, The Journal of the Acoustical Society of America.

[29]  Neil A. Macmillan,et al.  Detection Theory: A User's Guide , 1991 .

[30]  C E Rice,et al.  Human Echo Perception , 1967, Science.

[31]  Zhangli Chen,et al.  A new method of calculating auditory excitation patterns and loudness for steady sounds , 2011, Hearing Research.

[32]  B. Seeber,et al.  Dynamic-range compression affects the lateral position of sounds. , 2011, The Journal of the Acoustical Society of America.

[33]  Leslie R Bernstein,et al.  Sensitivity to interaural intensitive disparities: listeners' use of potential cues. , 2004, The Journal of the Acoustical Society of America.