Never-ending learning with dynamic hidden Markov network

Current automatic speech recognition systems have two distinctive modes of operation: training and recognition. After the training, system parameters are fixed, and if a mismatch between training and testing conditions occurs, an adaptation procedure is commonly applied. However, the adaptation methods change the system parameters in such a way that previously learned knowledge is irrecoverably destroyed. In searching for a solution to this problem and motivated by the results of recent neuro-biological studies, we have developed a network of hidden Markov states that is capable of unsupervised on-line adaptive learning while preserving the previously acquired knowledge. Speech patterns are represented by state sequences or paths through the network. The network can detect previously unseen patterns, and if such a new pattern is encountered, it is learned by adding new states and transitions to the network. Paths and states corresponding to spurious events or ”noises” and, therefore, rarely visited, are gradually removed. Thus, the network can grow and shrink when needed, i.e. it dynamically changes its structure. The learning process continues as long as the network lasts, i.e. theoretically forever, so it is called neverending learning. The output of the network is the best state sequence and the decoding is done concurrently with the learning. Thus the network always operates in a single learning/decoding mode. Initial experiments with a small database of isolated spelled letters showed that the Dynamic Hidden Markov network is indeed capable of never-ending learning and can perfectly recognize previously learned speech patterns. Index Terms: never-ending learning, life-long learning, dynamic hidden markov network, self-organization, topology representation.

[1]  Narendra Ahuja,et al.  A topological and temporal correlator network for spatiotemporal pattern learning, recognition, and recall , 1999, IEEE Trans. Neural Networks.

[2]  Dominique Béroule,et al.  An instance of coincidence detection architecture relying on temporal coding , 2004, IEEE Transactions on Neural Networks.

[3]  Stephen Grossberg,et al.  Nonlinear neural networks: Principles, mechanisms, and architectures , 1988, Neural Networks.

[4]  E. Capaldi,et al.  The organization of behavior. , 1992, Journal of applied behavior analysis.

[5]  Thomas Martinetz,et al.  Topology representing networks , 1994, Neural Networks.

[6]  Lawrence R. Rabiner,et al.  On integrating insights from human speech perception into automatic speech recognition , 2005, INTERSPEECH.

[7]  Stephen Grossberg,et al.  The ART of adaptive pattern recognition by a self-organizing neural network , 1988, Computer.

[8]  Shen Furao,et al.  An incremental network for on-line unsupervised classification and topology learning , 2006, Neural Networks.

[9]  R. French Catastrophic forgetting in connectionist networks , 1999, Trends in Cognitive Sciences.

[10]  Teuvo Kohonen,et al.  Self-Organization and Associative Memory , 1988 .

[11]  Shun-ichi Amari,et al.  Field theory of self-organizing neural nets , 1983, IEEE Transactions on Systems, Man, and Cybernetics.

[12]  Fred Henrik Hamker,et al.  Life-long learning Cell Structures--continuously learning without catastrophic interference , 2001, Neural Networks.

[13]  S. Kaplan The Physiology of Thought , 1950 .