Experimental Determination of Precision Requirements for Back-propagation Training of Artiicial Neural Networks

The impact of reduced weight and output precision on the back-propagation training algorithm Wer74, RHW86] is experimentally determined for a feed-forward multi-layer perceptron. In contrast with previous such studies, the network is large with over 20,000 weights, and is trained with a large, real-world data set of over 130,000 patterns to perform a diicult task, that of phoneme classiication for a continuous speech recognition system. The results indicate that 16b weight values are suucient to achieve training and classiication results comparable to 32b oating point, provided that weight and bias values are scaled separately, and that rounding rather than truncation is employed to reduce the precision of intermediary values. Output precision can be reduced to 8 bits without signiicant eeects on performance.

[1]  P. Werbos,et al.  Beyond Regression : "New Tools for Prediction and Analysis in the Behavioral Sciences , 1974 .

[2]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .

[3]  Tom Baker,et al.  Modifications to artificial neural networks models for Digital Hardware Implementation , 1988 .

[4]  H Hermansky,et al.  Perceptual linear predictive (PLP) analysis of speech. , 1990, The Journal of the Acoustical Society of America.

[5]  Hervé Bourlard,et al.  Continuous speech recognition using multilayer perceptrons with hidden Markov models , 1990, International Conference on Acoustics, Speech, and Signal Processing.

[6]  Eric Allman,et al.  RAP: a ring array processor for multilayer perceptron applications , 1990, International Conference on Acoustics, Speech, and Signal Processing.

[7]  Hynek Hermansky,et al.  Continuous speech recognition using PLP analysis with multilayer perceptrons , 1991, [Proceedings] ICASSP 91: 1991 International Conference on Acoustics, Speech, and Signal Processing.