A learning rule eliminating local minima in multilayer perceptrons

Convergence problems in the case of the generalized delta rule are discussed. On the basis of the analysis performed, a modification to the nonlinearity of processing elements is proposed; this modification is shown to smooth the cost function to be minimized during the learning phase. A variation to the generalized delta rule learning procedure, required by the introduced modification, is discussed. Extensive tests have been performed on several examples proposed in the technical literature. The tests show the effectiveness of the proposed procedure in improving the convergence properties of the backpropagation algorithm. In particular, it has been verified that the proposed modification virtually eliminates nonconvergence problems of a moderate eta value is used.<<ETX>>

[1]  R. Hecht-Nielsen,et al.  Back propagation error surfaces can have local minima , 1989, International 1989 Joint Conference on Neural Networks.

[2]  S. Y. Kung,et al.  An algebraic projection analysis for optimal hidden units size and learning rates in back-propagation learning , 1988, IEEE 1988 International Conference on Neural Networks.