Neural Network Learning from Ambiguous Training Data

After a brief review of the different types and causes of ambiguous training data and the problems of learning from such data, a class of multi-target models are presented which suggest that neural networks are even better at solving these problems than previously realized. They are able to learn which non-ambiguous subset of a larger ambiguous set of training data best captures any underlying regularities in that data and hence optimize generalization while minimizing the problems of overtraining. It is also shown how the deliberate generation of ambiguous training data can begin to solve some of the long-standing representational problems of mapping time sequences, such as the alignment problem for reading and spelling. The general ideas are illustrated throughout with the well- known problem of tex-to-phoneme conversion, and detailed results of a range of neural network simulations are presented.

[1]  R. Glushko The Organization and Activation of Orthographic Knowledge in Reading Aloud. , 1979 .

[2]  John A. Bullinaria Noise reduction by multi-target learning , 1994, ESANN.

[3]  Horacio Franco,et al.  Context-Dependent Multiple Distribution Phonetic Modeling with MLPs , 1992, NIPS.

[4]  Ehud D. Karnin,et al.  A simple procedure for pruning back-propagation trained neural networks , 1990, IEEE Trans. Neural Networks.

[5]  James D. Keeler,et al.  A Self-Organizing Integrated Segmentation and Recognition Neural Net , 1991, NIPS.

[6]  Geoffrey E. Hinton Learning Translation Invariant Recognition in Massively Parallel Networks , 1987, PARLE.

[7]  Terrence J. Sejnowski,et al.  Parallel Networks that Learn to Pronounce English Text , 1987, Complex Syst..

[8]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .

[9]  Hervé Bourlard,et al.  Generalization and Parameter Estimation in Feedforward Netws: Some Experiments , 1989, NIPS.

[10]  David E. Rumelhart,et al.  Predicting the Future: a Connectionist Approach , 1990, Int. J. Neural Syst..

[11]  Axel Cleeremans,et al.  Mechanisms of Implicit Learning: Connectionist Models of Sequence Processing , 1993 .

[12]  Anders Krogh,et al.  A Simple Weight Decay Can Improve Generalization , 1991, NIPS.

[13]  John A. Bullinaria,et al.  Connectionist Modelling of Spelling , 2019, Proceedings of the Sixteenth Annual Conference of the Cognitive Science Society.

[14]  Lawrence R. Rabiner,et al.  A tutorial on hidden Markov models and selected applications in speech recognition , 1989, Proc. IEEE.

[15]  Michael I. Jordan Attractor dynamics and parallelism in a connectionist sequential machine , 1990 .

[16]  D. Besner,et al.  Reading pseudohomophones: Implications for models of pronunciation assembly and the locus of word-frequency effects in naming. , 1987 .

[17]  James L. McClelland,et al.  A distributed, developmental model of word recognition and naming. , 1989, Psychological review.