Lexical boosting of noise-band speech in open- and closed-set formats

Abstract Recognition of spoken words in noise and in quiet is more accurate for Lexically Easy words (high frequency words with few similar-sounding neighbors) than for Lexically Hard words (low frequency words with many similar-sounding neighbors). Using monosyllables, the present set of two experiments extends this finding to a perceptually interesting class of stimuli and test formats. In both open- and closed-sets, listeners attempted to identify amplitude-modulated and bandpass-filtered words [Shannon, R., Zeng, F., Kamath, V., Wgonski, J., Ekelid, M., 1995. Speech recognition with primarily temporal cues. Science 270, 303–304]––noise-band speech––shown to simulate the performance of cochlear implant (CI) patients using the same number of frequency channels. The words were synthesized from a database that controls for Lexical Difficulty, Talker Identity and Talker Gender. Word recognition was significantly more accurate for Easy words in both the open- and the closed-set experiments. These results indicate that, even when spoken word recognition is challenged by noise-band speech, the Easy–Hard effect survives the perceptually uncertain conditions of word variability. Consequences for models of spoken word recognition are explored.

[1]  J A Bashford,et al.  Use of speech-modulated noise adds strong “bottom-up” cues for phonemic restoration , 1996, Perception & psychophysics.

[2]  W. Cooper,et al.  Sentence Processing: Psycholinguistic Studies Presented to Merrill Garrett. , 1980 .

[3]  A. Samuel Lexical uniqueness effects on phonemic restoration , 1987 .

[4]  J. D. Trout,et al.  Auditory and Visual Influences on Phonemic Restoration , 1990, Language and speech.

[5]  Michael Garman,et al.  Psycholinguistics: Accessing the mental lexicon , 1990 .

[6]  D. Pisoni,et al.  Speech perception without traditional speech cues. , 1981, Science.

[7]  M. Schroeder Reference Signal for Signal Quality Studies , 1968 .

[8]  Jacob Cohen Statistical Power Analysis for the Behavioral Sciences , 1969, The SAGE Encyclopedia of Research Design.

[9]  R V Shannon,et al.  Speech Recognition with Primarily Temporal Cues , 1995, Science.

[10]  H. Kucera,et al.  Computational analysis of present-day American English , 1967 .

[11]  D. Dirks,et al.  Examination of the Neighborhood Activation Theory in Normal and Hearing-Impaired Listeners , 2001, Ear and hearing.

[12]  P. Luce,et al.  A computational analysis of uniqueness points in auditory word recognition , 1986, Perception & psychophysics.

[13]  D J Van Tasell,et al.  Speech waveform envelope cues for consonant recognition. , 1987, The Journal of the Acoustical Society of America.

[14]  D. Pisoni,et al.  Recognition of spoken words by native and non-native listeners: talker-, listener-, and item-related factors. , 1999, The Journal of the Acoustical Society of America.

[15]  D B Pisoni,et al.  Some Considerations in Evaluating Spoken Word Recognition by Normal‐Hearing, Noise‐Masked Normal‐Hearing, and Cochlear Implant Listeners. I: The Effects of Response Format , 1997, Ear and hearing.

[16]  H. Nusbaum Sizing up the Hoosier Mental Lexicon: Measuring the Familiarity of 20,000 Words, Research on Speech Perception , 1984 .

[17]  R V Shannon,et al.  Speech recognition as a function of the number of electrodes used in the SPEAK cochlear implant speech processor. , 1997, Journal of speech, language, and hearing research : JSLHR.

[18]  M F Dorman,et al.  Recognition of Monosyllabic Words by Cochlear Implant Patients and by Normal-Hearing Subjects Listening to Words Processed through Cochlear Implant Signal Processing Strategies , 2000, The Annals of otology, rhinology & laryngology. Supplement.

[19]  D. Pisoni,et al.  Recognizing Spoken Words: The Neighborhood Activation Model , 1998, Ear and hearing.

[20]  William D Marslen-Wilson,et al.  Processing interactions and lexical access during word recognition in continuous speech , 1978, Cognitive Psychology.

[21]  D J Van Tasell,et al.  Temporal cues for consonant recognition: training, talker generalization, and use in evaluation of cochlear implants. , 1992, The Journal of the Acoustical Society of America.

[22]  P C Loizou,et al.  On the number of channels needed to understand speech. , 1999, The Journal of the Acoustical Society of America.

[23]  David B Pisoni,et al.  Individual Differences in Effectiveness of Cochlear Implants in Children Who Are Prelingually Deaf: New Process Measures of Performance. , 1999, The Volta review.

[24]  W A Dreschler,et al.  Importance of temporal-envelope cues in consonant recognition. , 1999, The Journal of the Acoustical Society of America.

[25]  R W Schvaneveldt,et al.  An activation--verification model for letter and word recognition: the word-superiority effect. , 1982, Psychological review.

[26]  Philipos C. Loizou,et al.  Word Recognition by Children Listening to Speech Processed into a Small Number of Channels: Data from Normal-Hearing Children and Children with Cochlear Implants , 2000, Ear and hearing.

[27]  M F Dorman,et al.  The Identification of Consonants and Vowels by Cochlear Implant Patients Using a 6‐Channel Continuous Interleaved Sampling Processor and by Normal‐Hearing Subjects Using Simulations of Processors with Two to Nine Channels , 1998, Ear and hearing.

[28]  M. Dorman,et al.  Speech intelligibility as a function of the number of channels of stimulation for signal processors using sine-wave and noise-band outputs. , 1997, The Journal of the Acoustical Society of America.

[29]  R. Shannon,et al.  Speech recognition in noise as a function of the number of spectral channels: comparison of acoustic hearing and cochlear implants. , 2001, The Journal of the Acoustical Society of America.