Discriminant clustering using an HMM isolated-word recognizer

One limitation of hidden Markov model (HMM) recognizers is that subword models are not learned but must be prespecified before training. This can lead to excessive computation during recognition and/or poor discrimination between similar sounding words. A training procedure called discriminant clustering is presented that creates subword models automatically. Node sequences from whole-word models are merged using statistical clustering techniques. This procedure reduced the computation required during recognition for a 35-word vocabulary by roughly one-third while maintaining a low error rate. It was also found that five iterations of the forward-backward algorithm are sufficient and that adding nodes to HMM word models improves performance until the minimum word transition time becomes excessive.<<ETX>>

[1]  L. Rabiner,et al.  An introduction to hidden Markov models , 1986, IEEE ASSP Magazine.

[2]  E. A. Martin,et al.  Multi-style training for robust isolated-word speech recognition , 1987, ICASSP '87. IEEE International Conference on Acoustics, Speech, and Signal Processing.

[3]  M. Tomlinson,et al.  The discriminative network: A mechanism for focusing recognition in whole-word pattern matching , 1983, ICASSP.

[4]  John A. Hartigan,et al.  Clustering Algorithms , 1975 .

[5]  Richard P. Lippmann,et al.  Two-stage discriminant analysis for improved isolated-word recognition , 1987, ICASSP '87. IEEE International Conference on Acoustics, Speech, and Signal Processing.

[6]  D. B. Paul,et al.  Speaker stress-resistant continuous speech recognition , 1988, ICASSP-88., International Conference on Acoustics, Speech, and Signal Processing.