Query-based learning for aerospace applications

Models of real-world applications often include a large number of parameters with a wide dynamic range, which contributes to the difficulties of neural network training. Creating the training data set for such applications becomes costly, if not impossible. In order to overcome the challenge, one can employ an active learning technique known as query-based learning (QBL) to add performance-critical data to the training set during the learning phase, thereby efficiently improving the overall learning/generalization. The performance-critical data can be obtained using an inverse mapping called network inversion (discrete network inversion and continuous network inversion) followed by oracle query. This paper investigates the use of both inversion techniques for QBL learning, and introduces an original heuristic to select the inversion target values for continuous network inversion method. Efficiency and generalization was further enhanced by employing node decoupled extended Kalman filter (NDEKF) training and a causality index (CI) as a means to reduce the input search dimensionality. The benefits of the overall QBL approach are experimentally demonstrated in two aerospace applications: a classification problem with large input space and a control distribution problem.

[1]  Lutz Prechelt,et al.  Automatic early stopping using cross validation: quantifying the criteria , 1998, Neural Networks.

[2]  David A. Cohn,et al.  Neural Network Exploration Using Optimal Experiment Design , 1993, NIPS.

[3]  Paul J. Werbos,et al.  Backpropagation Through Time: What It Does and How to Do It , 1990, Proc. IEEE.

[4]  David J. C. MacKay,et al.  Information-Based Objective Functions for Active Data Selection , 1992, Neural Computation.

[5]  David A. Cohn,et al.  Active Learning with Statistical Models , 1996, NIPS.

[6]  J. L. Vian,et al.  Neural Networks for Nonlinear Control Distribution , 1997 .

[7]  D Zipser,et al.  Learning the hidden structure of speech. , 1988, The Journal of the Acoustical Society of America.

[8]  Robert W. Blanning,et al.  An empirical measure of element contribution in neural networks , 1998, IEEE Trans. Syst. Man Cybern. Part C.

[9]  Seho Oh,et al.  Query based learning in a multilayered perceptron in the presence of data jitter , 1991, Proceedings of the First International Forum on Applications of Neural Networks to Power Systems.

[10]  A. Linden,et al.  Inversion of multilayer nets , 1989, International 1989 Joint Conference on Neural Networks.

[11]  S. Oh,et al.  An equivalence between sigmoidal gain scaling and training with noisy (jittered) input data , 1992, [Proceedings] 1992 RNNS/IEEE Symposium on Neuroinformatics and Neurocomputers.

[12]  Christian Cachin,et al.  Pedagogical pattern selection strategies , 1994, Neural Networks.

[13]  Kenji Fukumizu,et al.  Statistical active learning in multilayer perceptrons , 2000, IEEE Trans. Neural Networks Learn. Syst..

[14]  Kenji Baba,et al.  Explicit representation of knowledge acquired from plant historical data using neural network , 1990, 1990 IJCNN International Joint Conference on Neural Networks.

[15]  Simon Haykin,et al.  Neural Networks: A Comprehensive Foundation , 1998 .

[16]  David A. Cohn,et al.  Training Connectionist Networks with Queries and Selective Sampling , 1989, NIPS.

[17]  Dana Angluin,et al.  A Note on the Number of Queries Needed to Identify Regular Languages , 1981, Inf. Control..

[18]  D. Rumelhart,et al.  Generalization by weight-elimination applied to currency exchange rate prediction , 1991, [Proceedings] 1991 IEEE International Joint Conference on Neural Networks.

[19]  W. J. Studden,et al.  Theory Of Optimal Experiments , 1972 .

[20]  Robert J. Marks,et al.  Similarities of error regularization, sigmoid gain scaling, target smoothing, and training with jitter , 1995, IEEE Trans. Neural Networks.

[21]  Lee A. Feldkamp,et al.  Decoupled extended Kalman filter training of feedforward layered networks , 1991, IJCNN-91-Seattle International Joint Conference on Neural Networks.

[22]  P. Réfrégier,et al.  An Improved Version of the Pseudo-Inverse Solution for Classification and Neural Networks , 1989 .

[23]  Jenq-Neng Hwang,et al.  Interactive query learning for isolated speech recognition , 1992, Neural Networks for Signal Processing II Proceedings of the 1992 IEEE Workshop.

[24]  J. I. Minnix Fault tolerance of the backpropagation neural network trained on noisy inputs , 1992, [Proceedings 1992] IJCNN International Joint Conference on Neural Networks.

[25]  Jenq-Neng Hwang,et al.  Query-based learning applied to partially trained multilayer perceptrons , 1991, IEEE Trans. Neural Networks.

[26]  H. Guterman,et al.  Knowledge extraction from artificial neural network models , 1997, 1997 IEEE International Conference on Systems, Man, and Cybernetics. Computational Cybernetics and Simulation.

[27]  Sharad Singhal,et al.  Training Multilayer Perceptrons with the Extende Kalman Algorithm , 1988, NIPS.

[28]  Lee A. Feldkamp,et al.  Roles of learning rates, artificial process noise and square root filtering for extended Kalman filter training , 1999, IJCNN'99. International Joint Conference on Neural Networks. Proceedings (Cat. No.99CH36339).

[29]  Jenq-Neng Hwang,et al.  Query learning based on boundary search and gradient computation of trained multilayer perceptrons , 1990, 1990 IJCNN International Joint Conference on Neural Networks.

[30]  Jocelyn Sietsma,et al.  Creating artificial neural networks that generalize , 1991, Neural Networks.

[31]  Joel Ratsaby,et al.  Incremental Learning With Sample Queries , 1998, IEEE Trans. Pattern Anal. Mach. Intell..