We introduce the notion of suspect families of loading problems in an attempt of formalizing situations in which classical learning algorithms based on local optimization are likely to fail (because of local minima or numerical precision problems). We show that any loading problem belonging to a non-suspect family can be solved with optimal complexity by a canonical form of gradient descent with forced dynamics (i.e., for this class of problems no algorithm exhibits a better computational complexity than a slightly modified form of backpropagation). The analysis of this paper suggests intriguing links between the shape of the error surface attached to parametric learning systems (like neural networks) and the computational complexity of the corresponding optimization problem.
[1]
Amir F. Atiya,et al.
An accelerated learning algorithm for multilayer perceptron networks
,
1994,
IEEE Trans. Neural Networks.
[2]
Xiao-Hu Yu,et al.
Can backpropagation error surface not have local minima
,
1992,
IEEE Trans. Neural Networks.
[3]
Alberto Tesi,et al.
On the Problem of Local Minima in Backpropagation
,
1992,
IEEE Trans. Pattern Anal. Mach. Intell..
[4]
David Haussler,et al.
What Size Net Gives Valid Generalization?
,
1989,
Neural Computation.
[5]
J. J. Slotine,et al.
Tracking control of non-linear systems using sliding surfaces with application to robot manipulators
,
1983,
1983 American Control Conference.