The application of neural networks in pattern recognition problems constitute some difficult problems like initialization, choosing the correct architecture etc. In this paper the problem of ”how do networks find non-linear solutions” is addressed. By examining the properties of a simple (one input and one output) multi layered perceptron with sigmoidal transfer functions we try to show in what extend the network is able to find non-linear solutions starting in a linear initialization. In order to formalize the network’s behaviour, the discrimination capacity DC is introduced. This measure indicates whether the network has found any non-linear solutions. Experiments show that the values for DC that can be expected from the theory are indeed met but the mathematical analysis of the network is too complicated to exactly point out the exact points where specific changes of DC can be found.
[1]
David Haussler,et al.
What Size Net Gives Valid Generalization?
,
1989,
Neural Computation.
[2]
Vladimir Vapnik,et al.
Principles of Risk Minimization for Learning Theory
,
1991,
NIPS.
[3]
M. A. Kraaijveld.
Small Sample Behavior of Multi-Layer Feedforward Network Classifiers: Theoretical and Practical Aspects
,
1993
.
[4]
Robert P. W. Duin,et al.
The effective capacity of multilayer feedforward network classifiers
,
1994,
Proceedings of the 12th IAPR International Conference on Pattern Recognition, Vol. 3 - Conference C: Signal Processing (Cat. No.94CH3440-5).
[5]
Ida G. Sprinkhuizen-Kuyper,et al.
The Error Surface of the Simplest XOR Network Has Only Global Minima
,
1996,
Neural Computation.