Constraints on the processes responsible for the extrinsic normalization of vowels

Listeners tune in to talkers’ vowels through extrinsic normalization. We asked here whether this process could be based on compensation for the long-term average spectrum (LTAS) of preceding sounds and whether the mechanisms responsible for normalization are indifferent to the nature of those sounds. If so, normalization should apply to nonspeech stimuli. Previous findings were replicated with first-formant (F1) manipulations of speech. Targets on a [pt]–[pɛt] (low–high F1) continuum were labeled as [pt] more after high-F1 than after low-F1 precursors. Spectrally rotated nonspeech versions of these materials produced similar normalization. None occurred, however, with nonspeech stimuli that were less speechlike, even though precursor–target LTAS relations were equivalent to those used earlier. Additional experiments investigated the roles of pitch movement, amplitude variation, formant location, and the stimuli's perceived similarity to speech. It appears that normalization is not restricted to speech but that the nature of the preceding sounds does matter. Extrinsic normalization of vowels is due, at least in part, to an auditory process that may require familiarity with the spectrotemporal characteristics of speech.

[1]  D. Broadbent,et al.  Information Conveyed by Vowels , 1957 .

[2]  P. E. Stopp Frequency analysis and periodicity detection in hearing 1971, Plomp and Smoorenburg (Editors). Leiden, Netherlands: Sijthoff Leiden. Cloth, Fl. 60 , 1971 .

[3]  S. S. Stevens Frequency Analysis and Periodicity Detection in Hearing. , 1972 .

[4]  E. Owens,et al.  An Introduction to the Psychology of Hearing , 1997 .

[5]  B. Moore An Introduction to the Psychology of Hearing , 1977 .

[6]  R E Remez,et al.  Perceptual normalization of vowels produced by sinusoidal voices. , 1987, Journal of experimental psychology. Human perception and performance.

[7]  T. M. Nearey Static, dynamic, and relational properties in vowel perception. , 1989, The Journal of the Acoustical Society of America.

[8]  B. Moore An introduction to the psychology of hearing, 3rd ed. , 1989 .

[9]  A. J. Watkins Central, auditory mechanisms of perceptual compensation for spectral-envelope distortion. , 1991, The Journal of the Acoustical Society of America.

[10]  A. J. Watkins,et al.  Perceptual compensation for speaker differences and for spectral-envelope distortion. , 1994, The Journal of the Acoustical Society of America.

[11]  A. J. Watkins,et al.  Effects of spectral contrast on perceptual compensation for spectral-envelope distortion. , 1996, The Journal of the Acoustical Society of America.

[12]  A. Lotto,et al.  Perceptual compensation for coarticulation by Japanese quail (Coturnix coturnix japonica). , 1997, The Journal of the Acoustical Society of America.

[13]  Elizabeth A. Strand,et al.  Auditory–visual integration of talker gender in vowel perception , 1999 .

[14]  A. Lotto,et al.  Behavioral examinations of the level of auditory processing of speech context effects , 2002, Hearing Research.

[15]  Sarah C. Sullivan,et al.  Central locus for nonspeech context effects on phonetic identification. , 2003, The Journal of the Acoustical Society of America.

[16]  Michael Kiefte,et al.  Sensitivity to change in perception of speech , 2003, Speech Commun..

[17]  Roel Smits,et al.  An acoustic description of the vowels of Northern and Southern Standard Dutch. , 2004, The Journal of the Acoustical Society of America.

[18]  L. Holt Temporally Nonadjacent Nonlinguistic Sounds Affect Speech Categorization , 2005, Psychological science.

[19]  Michael Kiefte,et al.  The relative importance of spectral tilt in monophthongs and diphthongs. , 2005, The Journal of the Acoustical Society of America.

[20]  Michael Kiefte,et al.  Chapter 6 – Speech Perception within a Biologically Realistic Information-Theoretic Framework , 2006 .

[21]  Holger Mitterer,et al.  On the causes of compensation for coarticulation: Evidence for phonological mediation , 2006, Perception & psychophysics.

[22]  L. Holt The mean matters: effects of statistically defined nonspeech spectral distributions on speech categorization. , 2006, The Journal of the Acoustical Society of America.

[23]  G. Altmann,et al.  The Oxford Handbook of Psycholinguistics , 2007 .

[24]  Dominic W. Massaro,et al.  Audiovisual speech perception and word recognition , 2007 .

[25]  Michael Kiefte,et al.  Absorption of reliable spectral characteristics in auditory perception. , 2008, The Journal of the Acoustical Society of America.

[26]  K. Kluender Speech perception within a biologically realistic information‐theoretic framework. , 2008 .

[27]  A. Samuel,et al.  Perceptual learning for speech , 2009, Attention, perception & psychophysics.

[28]  Christian E Stilp,et al.  Auditory color constancy: Calibration to reliable spectral properties across nonspeech context and targets , 2010, Attention, perception & psychophysics.

[29]  E Klu Auditory color constancy: Calibration to reliable spectral properties across nonspeech context and targets , 2010 .

[30]  Michael J Cortese,et al.  Handbook of Psycholinguistics , 2011 .

[31]  Holger Mitterer,et al.  Recognizing reduced forms: Different processing mechanisms for similar reductions , 2011, J. Phonetics.

[32]  Hwee Ling Lee Audiovisual speech perception , 2012 .