Nonlinear Autoassociation Is Not Equivalent to PCA

A common misperception within the neural network community is that even with nonlinearities in their hidden layer, autoassociators trained with backpropagation are equivalent to linear methods such as principal component analysis (PCA). Our purpose is to demonstrate that nonlinear autoassociators actually behave differently from linear methods and that they can outperform these methods when used for latent extraction, projection, and classification. While linear autoassociators emulate PCA, and thus exhibit a flat or unimodal reconstruction error surface, autoassociators with nonlinearities in their hidden layer learn domains by building error reconstruction surfaces that, depending on the task, contain multiple local valleys. This interpolation bias allows nonlinear autoassociators to represent appropriate classifications of nonlinear multimodal domains, in contrast to linear autoassociators, which are inappropriate for such tasks. In fact, autoassociators with hidden unit nonlinearities can be shown to perform nonlinear classification and nonlinear recognition.

[1]  Erkki Oja,et al.  Subspace methods of pattern recognition , 1983 .

[2]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .

[3]  Paul W. Munro,et al.  Principal Components Analysis Of Images Via Back Propagation , 1988, Other Conferences.

[4]  N. E. Sharkey,et al.  Models of cognition : a review of cognitive science , 1989 .

[5]  Kurt Hornik,et al.  Neural networks and principal component analysis: Learning from examples without local minima , 1989, Neural Networks.

[6]  Leonid Kruglyak,et al.  How to Solve the N Bit Encoder Problem with Just Two Hidden Units , 1990, Neural Computation.

[7]  Anders Krogh,et al.  Introduction to the theory of neural computation , 1994, The advanced book program.

[8]  Dhananjay S. Phatak,et al.  Construction of Minimal n-2-n Encoders for Any n , 1993, Neural Computation.

[9]  Nathalie Japkowicz,et al.  A Novelty Detection Approach to Classification , 1995, IJCAI.

[10]  Xun Liang,et al.  How to solve the N-bit encoder problem with one hidden unit and polynomially increasing weights and thresholds , 1995, Neurocomputing.

[11]  Stephen Jose Hanson,et al.  A Neural Network Autoassociator for Induction Motor Failure Prediction , 1995, NIPS.

[12]  Nathalie Japkowicz,et al.  Concept learning in the absence of counterexamples: an autoassociation-based approach to classification , 1999 .

[13]  D. Mareschal,et al.  Modeling Infant Speech Sound Discrimination Using Simple Associative Networks. , 2001, Infancy : the official journal of the International Society on Infant Studies.