A Modified Error Function to Improve the Error Back-Propagation Algorithm for Multi-Layer Perceptrons

This paper proposes a modified error function to improve the error back-propagation (EBP) algorithm for multi-Layer perceptrons (MLPs) which suffers from slow learning speed. It can also suppress over-specialization for training patterns that occurs in an algorithm based on a cross-entropy cost function which markedly reduces learning time. In the similar way as the cross-entropy function, our new function accelerates the learning speed of the EBP algorithm by allowing the output node of the MLP to generate a strong error signal when the output node is far from the desired value. Moreover, it prevents the overspecialization of learning for training patterns by letting the output node, whose value is close to the desired value, generate a weak error signal. In a simulation study to classify handwritten digits in the CEDAR [1] database, the proposed method attained 100% correct classification for the training patterns after only 50 sweeps of learning, while the original EBP attained only 98.8% after 500 sweeps. Also, our method shows mean-squared error of 0.627 for the test patterns, which is superior to the error 0.667 in the cross-entropy method. These results demonstrate that our new method excels others in learning speed as well as in generalization.

[1]  Sang-Hoon Oh,et al.  An analysis of premature saturation in back propagation learning , 1993, Neural Networks.

[2]  R. K. M. Cheung,et al.  Relative effectiveness of training set patterns for backpropagation , 1990, 1990 IJCNN International Joint Conference on Neural Networks.

[3]  Richard Lippmann,et al.  Neural Network Classifiers Estimate Bayesian a posteriori Probabilities , 1991, Neural Computation.

[4]  Jonathan J. Hull,et al.  A Database for Handwritten Text Recognition Research , 1994, IEEE Trans. Pattern Anal. Mach. Intell..

[5]  Arjen van Ooyen,et al.  Improving the convergence of the back-propagation algorithm , 1992, Neural Networks.

[6]  N. Tepedelenlioglu,et al.  The effect of the activation function of the back propagation algorithm , 1989, IEEE 1989 International Conference on Systems Engineering.

[7]  Brian A. Telfer,et al.  Energy functions for minimizing misclassification error with minimum-complexity networks , 1994, Neural Networks.

[8]  Jin Wang,et al.  Weight smoothing to improve network generalization , 1994, IEEE Trans. Neural Networks.

[9]  Soo-Young Lee,et al.  Hybrid Hebbian / Back-Propagation Learning Rule for Improved Generalization of Multilayer Feed-Forward Neural Networks , 1994 .

[10]  Ryotaro Kamimura,et al.  Maximum Entropy Principle : Improving Generalization Performance by Maximizing the Number of Internal Representations , 1994 .

[11]  Adam Krzyzak,et al.  Classification of large set of handwritten characters using modified back propagation model , 1990, 1990 IJCNN International Joint Conference on Neural Networks.