Level-weighted averaging in elevation to synchronous amplitude-modulated sounds

To program a goal-directed response in the presence of multiple sounds, the audiomotor system should separate the sound sources. The authors examined whether the brain can segregate synchronous broadband sounds in the midsagittal plane, using amplitude modulations as an acoustic discrimination cue. To succeed in this task, the brain has to use pinna-induced spectral-shape cues and temporal envelope information. The authors tested spatial segregation performance in the midsagittal plane in two paradigms in which human listeners were required to localize, or distinguish, a target amplitude-modulated broadband sound when a non-modulated broadband distractor was played simultaneously at another location. The level difference between the amplitude-modulated and distractor stimuli was systematically varied, as well as the modulation frequency of the target sound. The authors found that participants were unable to segregate, or localize, the synchronous sounds. Instead, they invariably responded toward a level-weighted average of both sound locations, irrespective of the modulation frequency. An increased variance in the response distributions for double sounds of equal level was also observed, which cannot be accounted for by a segregation model, or by a probabilistic averaging model.To program a goal-directed response in the presence of multiple sounds, the audiomotor system should separate the sound sources. The authors examined whether the brain can segregate synchronous broadband sounds in the midsagittal plane, using amplitude modulations as an acoustic discrimination cue. To succeed in this task, the brain has to use pinna-induced spectral-shape cues and temporal envelope information. The authors tested spatial segregation performance in the midsagittal plane in two paradigms in which human listeners were required to localize, or distinguish, a target amplitude-modulated broadband sound when a non-modulated broadband distractor was played simultaneously at another location. The level difference between the amplitude-modulated and distractor stimuli was systematically varied, as well as the modulation frequency of the target sound. The authors found that participants were unable to segregate, or localize, the synchronous sounds. Instead, they invariably responded toward a level-w...

[1]  C J Darwin,et al.  Listening to speech in the presence of other sounds , 2008, Philosophical Transactions of the Royal Society B: Biological Sciences.

[2]  P M Hofman,et al.  Spectro-temporal factors in two-dimensional human sound localization. , 1998, The Journal of the Acoustical Society of America.

[3]  Albert S. Bregman,et al.  The Auditory Scene. (Book Reviews: Auditory Scene Analysis. The Perceptual Organization of Sound.) , 1990 .

[4]  D. M. Green,et al.  Sound localization by human listeners. , 1991, Annual review of psychology.

[5]  Guy J. Brown,et al.  Speech segregation based on sound localization , 2003 .

[6]  A John Van Opstal,et al.  The influence of duration and level on human sound localization. , 2004, The Journal of the Acoustical Society of America.

[7]  Franz J. Neyer,et al.  A Gentle Introduction to Bayesian Analysis: Applications to Developmental Research , 2013, Child development.

[8]  Virginia Best,et al.  Separation of concurrent broadband sound sources by human listeners. , 2004, The Journal of the Acoustical Society of America.

[9]  F L Wightman,et al.  Headphone simulation of free-field listening. II: Psychophysical validation. , 1989, The Journal of the Acoustical Society of America.

[10]  Josh H. McDermott The cocktail party problem , 2009, Current Biology.

[11]  Paul M. Hofman,et al.  Relearning sound localization with new ears , 1998, Nature Neuroscience.

[12]  Barak A. Pearlmutter,et al.  Sparse Representations for the Cocktail Party Problem , 2006, The Journal of Neuroscience.

[13]  William A. Yost,et al.  Localizing the sources of two independent noises: role of time varying amplitude differences. , 2013, The Journal of the Acoustical Society of America.

[14]  F. Ottes,et al.  Metrics of saccade responses to visual double stimuli: Two different modes , 1984, Vision Research.

[15]  A. John Van Opstal,et al.  Contribution of Head Shadow and Pinna Cues to Chronic Monaural Sound Localization , 2004 .

[16]  Masakazu Konishi,et al.  Mechanisms of sound localization in the barn owl (Tyto alba) , 1979, Journal of comparative physiology.

[17]  E. C. Cherry Some Experiments on the Recognition of Speech, with One and with Two Ears , 1953 .

[18]  Frank Jäkel,et al.  Bayesian inference for psychometric functions. , 2005, Journal of vision.

[19]  A. V. van Opstal,et al.  Pinna Cues Determine Orienting Response Modes to Synchronous Sounds in Elevation , 2010, The Journal of Neuroscience.

[20]  Eric-Jan Wagenmakers,et al.  An encompassing prior generalization of the Savage-Dickey density ratio , 2010, Comput. Stat. Data Anal..

[21]  Terrence J. Sejnowski,et al.  An Information-Maximization Approach to Blind Separation and Blind Deconvolution , 1995, Neural Computation.

[22]  John K Kruschke,et al.  Bayesian data analysis. , 2010, Wiley interdisciplinary reviews. Cognitive science.

[23]  Jeffrey S. Johnson,et al.  Segregating two simultaneous sounds in elevation using temporal envelope: Human psychophysics and a physiological model. , 2015, The Journal of the Acoustical Society of America.

[24]  Douglas P. Munoz,et al.  The effect of spatial–temporal audiovisual disparities on saccades in a complex scene , 2009, Experimental Brain Research.

[25]  D. Burr,et al.  The Ventriloquist Effect Results from Near-Optimal Bimodal Integration , 2004, Current Biology.

[26]  D. Robinson,et al.  A METHOD OF MEASURING EYE MOVEMENT USING A SCLERAL SEARCH COIL IN A MAGNETIC FIELD. , 1963, IEEE transactions on bio-medical engineering.

[27]  Barbara Shinn-Cunningham,et al.  Spatial cues alone produce inaccurate sound segregation: the effect of interaural time differences. , 2012, The Journal of the Acoustical Society of America.

[28]  J. Findlay Global visual processing for saccadic eye movements , 1982, Vision Research.

[29]  J. C. Middlebrooks,et al.  Weighting of Spatial and Spectro-Temporal Cues for Auditory Scene Analysis by Human Listeners , 2013, PloS one.

[30]  J. Dickey The Weighted Likelihood Ratio, Linear Hypotheses on Normal Location Parameters , 1971 .