Effectiveness of feature extraction in neural network architectures for novelty detection

This paper examines the performance of seven neural network architectures in classifying and detecting novel events contained within data collected from turbine sensors. Several different multilayer perceptrons were built and trained using backpropagation, conjugate gradient and quasi-Newton training algorithms. In addition, linear networks, radial basis function networks, probabilistic networks and Kohonen self organising feature maps were also built and trained, with the objective of discovering the most appropriate architecture. Because of the large input set involved in practice, feature extraction is examined to reduce the input features, the techniques considered being stepwise linear regression and a genetic algorithm. The results of these experiments have demonstrated an improvement in classification performance for multilayer perceptrons, Kohonen and probabilistic networks, using both genetic algorithms and stepwise linear regression over other architectures considered in this work. In addition, linear regression also performed better than a genetic algorithm for feature extraction. For classification problems involving a clear two class structure we consider a synthesis of stepwise linear regression with any of the architectures listed above to offer demonstrable improvements in performance for important real world tasks.

[1]  Lionel Tarassenko,et al.  Choosing an appropriate model for novelty detection , 1997 .

[2]  Robert Milne,et al.  Diagnosis of Dynamic Systems Based on Explicit and Implicit Behavioural Models: An Application to Gas Turbines in Esprit Project TIGER , 1995, SCAI.

[3]  Shahla Keyvan,et al.  Application of artificial neural networks for the development of a signal monitoring system , 1997 .

[4]  Heekuck Oh,et al.  Neural Networks for Pattern Recognition , 1993, Adv. Comput..

[5]  Belle R. Upadhyaya,et al.  An integrated signal processing and neural networks system for steam generator tubing diagnostics using eddy current inspection , 1996 .

[6]  David E. Goldberg,et al.  Genetic Algorithms in Search Optimization and Machine Learning , 1988 .

[7]  M. Hestenes,et al.  Methods of conjugate gradients for solving linear systems , 1952 .

[8]  Wray L. Buntine Theory Refinement on Bayesian Networks , 1991, UAI.

[9]  B. AfeArd CALCULATING THE SINGULAR VALUES AND PSEUDOINVERSE OF A MATRIX , 2022 .

[10]  David Lowe,et al.  Radial basis function networks , 1998 .

[11]  Stefan Wermter,et al.  Hybrid neural systems: from simple coupling to fully integrated neural networks , 1999 .

[12]  Simon Haykin,et al.  Neural Networks: A Comprehensive Foundation , 1998 .

[13]  John H. Holland,et al.  Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence , 1992 .

[14]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .

[15]  Etienne Barnard,et al.  Optimization for training neural nets , 1992, IEEE Trans. Neural Networks.

[16]  Elizabeth A. Peck,et al.  Introduction to Linear Regression Analysis , 2001 .

[17]  A. L. Edwards,et al.  An introduction to linear regression and correlation. , 1985 .

[18]  Donald F. Specht,et al.  Probabilistic neural networks , 1990, Neural Networks.