A mixture linear model with target-directed dynamics for spontaneous speech recognition

In this paper, a mixture linear dynamic model (MLDM) for speech recognition is developed and evaluated, where several linear dynamic systems are combined (mixed) to represent different vocaltract-resonance (VTR) dynamic behavior and the mapping relationships between the VTRs and the acoustic observation. Each linear dynamic model is formulated as a stale-space system, where the VTR's target-directed dynamic property is incorporated in the state equation and a linear regression function is used for the observation equation to piecewise linearly approximate the nonlinear mapping relationship. A version of the generalized EM algorithm is developed for learning the model parameters, where the VTR targets are constrained to change only at the segmental level (rather than at the frame level) in the parameter learning and model scoring algorithms. Speech recognition experiments are carried out to evaluate this new model using the N-best re-scoring paradigm in a Switchboard task. Compared with a baseline recognizer using the triphone HMM acoustic model, the new recognizer demonstrates superior performance under a number of experimental conditions.