Supervised and Unsupervised Learning in Linear Networks

We give an overview of the main facts on supervised and unsupervised learning in networks of linear units and present several new results and open questions. In the case of back-propagation, the complete structure of the landscape of the error function and its connections to known statistical techniques such as linear regression, principal component analysis and discriminant analysis have been established. Here, we examine the dynamical aspects of the learning process, how in certain cases the spectral properties of a covariance matrix are learnt according to the order defined by the eigenvalues, and the effects of noise. In the low noise limit, we prove that the strategy adopted by the networks are unchanged whereas in the high noise limit, the solution adopted is one of complete redundancy. In the case of unsupervised learning, several algorithms based on various hebbian and anti-hebbian mechanisms are reviewed together with the structure of their fixed points. We show that three “symmetric” algorithms suggested in the literature (Oja, 1982; Williams, 1985; Baldi, 1988) are in fact equivalent. Results of simulations are presented.

[1]  Ralph Linsker,et al.  Self-organization in a perceptual network , 1988, Computer.

[2]  Kurt Hornik,et al.  Neural networks and principal component analysis: Learning from examples without local minima , 1989, Neural Networks.

[3]  E. Oja Simplified neuron model as a principal component analyzer , 1982, Journal of mathematical biology.

[4]  Pierre Baldi,et al.  Linear Learning: Landscapes and Algorithms , 1988, NIPS.

[5]  Patrick Gallinari,et al.  Multilayer perceptrons and data analysis , 1988, IEEE 1988 International Conference on Neural Networks.

[6]  Y. Chauvin,et al.  Principal component analysis by gradient descent on a constrained linear Hebbian cell , 1989, International 1989 Joint Conference on Neural Networks.