Shared Weights Neural Networks in Image Analysis

This thesis is concerned with the use of shared weights neural networks in image analysis. This type of neural network has been extensively described in literature since 1989. It is believed that networks incorporating shared weights are capable of local, shift-invariant feature extraction due to the restrictions placed on their architecture. The rst experiments were focused mainly on the neural network architectures as suggested by, amongst others, Le Cun et al. LBD + 89, LBD + 90, LJB + 89] and Viennet Vie93]. These archi-tectures basically are back-propagation neural networks. However, they restrain the number of free parameters and introduce the notion of receptive elds, combining local information into more abstract patterns at a higher level. Three of these networks were tested on the problem of handwritten digit recognition and the results were compared to those of methods based on other feature extraction or classiication techniques. As an intermezzo, a second order minimization method used in neural network training will be discussed, the pseudo-Newton method. There has been some debate on the advantages and disadvantages of this method. Experiments have been performed on a simulated error surface and on real neural network problems to investigate this method's behaviour. The goal of the last part of the thesis is to place the speciic example of handwritten digit recognition using shared weights neural networks in a more general context. The feature extraction capabilities of these networks were investigated on some problems and the learned templates and lters were compared to classic image processing approaches. Some qualitative conclusions can be drawn from these experiments concerning the innuence of network and data-set complexity, training algorithms and their parameters and weight initialization on training results. v vi Acknowledgements This thesis was written as a report of the research I performed for my graduation in the Know-like to thank them both for their guidance, help and support, and for showing me new ways to look at artiicial intelligence and pattern recognition. It seems to have become a tradition for students of the Pattern Recogniton Group to thank just about every person they have met in Delft in their M.Sc. theses. Following that tradition, I would simply like to thank everyone who would like to be thanked. Special thanks go to ir. Michael van Ginkel for (too) frequently letting me pick his brain on some image processing problems. vii viii Contents Abstract v Acknowledgements vii 1 Introduction …

[1]  Terrence J. Sejnowski,et al.  Parallel Networks that Learn to Pronounce English Text , 1987, Complex Syst..

[2]  S. Ragazzini,et al.  Learning of word stress in a sub-optimal second order back-propagation neural network , 1988, IEEE 1988 International Conference on Neural Networks.

[3]  Françoise Fogelman-Soulié,et al.  Multi-Modular Neural Network Architectures: Applications in Optical Character and Human Face Recognition , 1993, Int. J. Pattern Recognit. Artif. Intell..

[4]  Lawrence D. Jackel,et al.  Handwritten character recognition using neural network architectures , 1990 .

[5]  Kunihiko Fukushima,et al.  Neocognitron: A new algorithm for pattern recognition tolerant of deformations and shifts in position , 1982, Pattern Recognit..

[6]  Yann LeCun,et al.  Constrained neural networks for pattern recognition , 1991 .

[7]  Yoshua Bengio,et al.  Neural networks for speech and sequence recognition , 1996 .

[8]  Lawrence D. Jackel,et al.  Backpropagation Applied to Handwritten Zip Code Recognition , 1989, Neural Computation.

[9]  S. Makram-Ebeid,et al.  A rationalized error back-propagation learning algorithm , 1989, International 1989 Joint Conference on Neural Networks.

[10]  William H. Press,et al.  Numerical recipes in C , 2002 .

[11]  Yann LeCun,et al.  Improving the convergence of back-propagation learning with second-order methods , 1989 .

[12]  I. Guyon,et al.  Handwritten digit recognition: applications of neural network chips and automatic learning , 1989, IEEE Communications Magazine.

[13]  Emmanuel Viennet Architectures connexionnistes multi-modulaires. Application a l'analyse de scenes , 1993 .

[14]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .

[15]  Conclusions and Recommendations 7. Conclusions and Recommendations , 2022 .

[16]  Azriel Rosenfeld,et al.  Digital Picture Processing , 1976 .

[17]  Lawrence D. Jackel,et al.  Handwritten Digit Recognition with a Back-Propagation Network , 1989, NIPS.

[18]  Thomas Jackson,et al.  Neural Computing - An Introduction , 1990 .

[19]  Isabelle Guyon,et al.  Capacity control in linear classifiers for pattern recognition , 1992, Proceedings., 11th IAPR International Conference on Pattern Recognition. Vol.II. Conference B: Pattern Recognition Methodology and Systems.

[20]  Alice J. O'Toole,et al.  Neural computing: Theory and practice: Philip D. Wasserman, Van Nostrand Reinhold: New York, 1989, $36.95, 230 pp. ISBN 442-207-433 , 1990 .