Effects of reverberation on perceptual segregation of competing voices.

Two experiments investigated the effect of reverberation on listeners' ability to perceptually segregate two competing voices. Culling et al. [Speech Commun. 14, 71-96 (1994)] found that for competing synthetic vowels, masked identification thresholds were increased by reverberation only when combined with modulation of fundamental frequency (F0). The present investigation extended this finding to running speech. Speech reception thresholds (SRTs) were measured for a male voice against a single interfering female voice within a virtual room with controlled reverberation. The two voices were either (1) co-located in virtual space at 0 degrees azimuth or (2) separately located at +/-60 degrees azimuth. In experiment 1, target and interfering voices were either normally intonated or resynthesized with a fixed F0. In anechoic conditions, SRTs were lower for normally intonated and for spatially separated sources, while, in reverberant conditions, the SRTs were all the same. In experiment 2, additional conditions employed inverted F0 contours. Inverted F0 contours yielded higher SRTs in all conditions, regardless of reverberation. The results suggest that reverberation can seriously impair listeners' ability to exploit differences in F0 and spatial location between competing voices. The levels of reverberation employed had no effect on speech intelligibility in quiet.

[1]  IEEE Recommended Practice for Speech Quality Measurements , 1969, IEEE Transactions on Audio and Electroacoustics.

[2]  Jont B. Allen,et al.  Image method for efficiently simulating small‐room acoustics , 1976 .

[3]  T. W. Parsons Separation of speech from interfering speech by means of harmonic selection , 1976 .

[4]  A. M. Mimpen,et al.  Improving the reliability of testing the speech reception threshold for sentences. , 1979, Audiology : official organ of the International Society of Audiology.

[5]  C J Darwin,et al.  Phoneme-monitoring reaction time and preceding prosody: Effects of stop closure duration and of fundamental frequency , 1981, Perception & psychophysics.

[6]  S. G. Nooteboom,et al.  Intonation and the perceptual separation of simultaneous voices , 1982 .

[7]  D. O'Shaughnessy,et al.  Linguistic modality effects on fundamental frequency in speech. , 1983, The Journal of the Acoustical Society of America.

[8]  Michaël Titus Maria Scheffers,et al.  Sifting vowels. Auditory pitch analysis and sound segregation. , 1983 .

[9]  Anne Cutler,et al.  The use of prosodic information in word recognition , 1984 .

[10]  T. Houtgast,et al.  A review of the MTF concept in room acoustics and its use for estimating speech intelligibility in auditoria , 1985 .

[11]  P. Peterson Simulating the response of multiple microphones to a single acoustic source in a reverberant room. , 1986, The Journal of the Acoustical Society of America.

[12]  R. Plomp A signal-to-noise ratio model for the speech-reception threshold of the hearing impaired. , 1986, Journal of speech and hearing research.

[13]  Anne Cutler,et al.  Components of prosodic effects in speech recognition , 1987 .

[14]  John F. Culling,et al.  Speech perception seen through the ear , 1989, Speech Commun..

[15]  John F. Culling,et al.  Periodicity of maskers not targets determines ease of perceptual segregation using differences in fundamental frequency , 1992 .

[16]  C. Darwin,et al.  Perceptual separation of simultaneous vowels: within and across-formant grouping by F0. , 1993, The Journal of the Acoustical Society of America.

[17]  C. Darwin,et al.  The role of timbre in the segregation of simultaneous voices with intersecting F0 contours , 1993, Perception & psychophysics.

[18]  John F. Culling,et al.  Effects of simulated reverberation on the use of binaural cues and fundamental-frequency differences for separating concurrent vowels , 1994, Speech Commun..

[19]  John F. Culling Signal-processing software for teaching and research in psychoacoustics under UNIX and X-Windows , 1996 .

[20]  C. M. Marin,et al.  Concurrent vowel identification II: Effects of phase, harmonicity and task , 1997 .

[21]  A. Cheveigné Concurrent vowel identification. III. A neural model of harmonic interference cancellation , 1997 .

[22]  Ann Cutler,et al.  Prosody in the Comprehension of Spoken Language: A Literature Review , 1997, Language and speech.

[23]  P F Assmann,et al.  Pitches of concurrent vowels. , 1994, The Journal of the Acoustical Society of America.

[24]  H S Colburn,et al.  Binaural sluggishness in the perception of tone sequences and speech in noise. , 2000, The Journal of the Acoustical Society of America.

[25]  R W Hukin,et al.  Effects of reverberation on spatial, prosodic, and vocal-tract size cues to selective attention. , 2000, The Journal of the Acoustical Society of America.