Diagnosis using backpropagation neural networks—analysis and criticism

Abstract Artificial neural networks based on a feedforward architecture and trained by the backpropagation technique have recently been applied to static fault diagnosis problems. The networks are used to classify measurement vectors into a set of predefined categories that represent the various functional and malfunctional states of the process. While the networks can usually produce decision surfaces that correctly classify the training examples, regions of the input space not occupied by training data are classified arbitrarily. As a result, the networks may not accurately extrapolate from the training data. Although extrapolation is not required under ideal circumstances, in practice the network may be required to extrapolate when undersized training sets are used, when parent distributions of fault classes undergo shifts subsequent to training, and when the input data is corrupted by missing or biased sensors. These situations cause relatively high error rates for the neural classifier. A related probem is that the networks cannot detect when they lack the data for a reliable classification, a serious deficiency in many practical applications. Classifiers based on distance metrics assign regions of the input space according to their proximity to the training data, and thus extrapolation is not arbitrary but based on the most relevant data. Distance-based classifiers perform better under nonideal conditions and are to be preferred to neural network classifiers in diagnostic applications.

[1]  Venkat Venkatasubramanian,et al.  A neural network methodology for process fault diagnosis , 1989 .

[2]  Ramaswamy Vaidyanathan,et al.  Process fault detection and diagnosis using neural networks , 1990 .

[3]  George Cybenko,et al.  Approximation by superpositions of a sigmoidal function , 1992, Math. Control. Signals Syst..

[4]  Masahiro Abe,et al.  Incipient fault diagnosis of chemical processes via artificial neural networks , 1989 .

[5]  David Haussler,et al.  What Size Net Gives Valid Generalization? , 1989, Neural Computation.

[6]  Mark A. Kramer,et al.  Improvement of the backpropagation algorithm for training neural networks , 1990 .

[7]  Peter E. Hart,et al.  Nearest neighbor pattern classification , 1967, IEEE Trans. Inf. Theory.

[8]  Terrence J. Sejnowski,et al.  Analysis of hidden units in a layered network trained to classify sonar targets , 1988, Neural Networks.

[9]  Josiah C. Hoskins,et al.  Artificial neural network models for knowledge representation in chemical engineering , 1990 .

[10]  Lyle H. Ungar,et al.  Adaptive networks for fault diagnosis and process control , 1990 .

[11]  Richard Lippmann,et al.  Practical Characteristics of Neural Network and Conventional Pattern Classifiers on Artificial and Speech Problems , 1989, NIPS.

[12]  John Moody,et al.  Fast Learning in Networks of Locally-Tuned Processing Units , 1989, Neural Computation.

[13]  J. Makhoul,et al.  Formation of disconnected decision regions with a single hidden layer , 1989, International 1989 Joint Conference on Neural Networks.

[14]  Richard Lippmann,et al.  Review of Neural Networks for Speech Recognition , 1989, Neural Computation.

[15]  J. J. Ferrada,et al.  Application of neural networks for fault diagnosis in plant production , 1989 .

[16]  S. Naidu,et al.  Application of Neural Networks on the Detection of Sensor Failure During the Operation of a Control System , 1989, 1989 American Control Conference.