Learnt Topology Gating Artificial Neural Networks

This work combines several established regression and meta-learning techniques to give a holistic regression model and presents the proposed learnt topology gating artificial neural networks (LTGANN) model in the context of a general architecture previously published by the authors. The applied regression techniques are artificial neural networks, which are on one hand used as local experts for the regression modelling and on the other hand as gating networks. The role of the gating networks is to estimate the prediction error of the local experts dependent on the input data samples. This is achieved by relating the input data space to the performance of the local experts, and thus building a performance map, for each of the local experts. The estimation of the prediction error is then used for the weighting of the local experts predictions. Another advantage of our approach is that the particular neural networks are unconstrained in terms of the number of hidden units. It is only necessary to define the range within which the number of hidden units has to be generated. The model links the topology to the performance, which has been achieved by the network with the given complexity, using a probabilistic approach. As the model was developed in the context of process industry data, it is evaluated using two industrial data sets. The evaluation has shown a clear advantage when using a model combination and meta-learning approach as well as demonstrating the higher performance of LTGANN when compared to a standard combination method.

[1]  Ricardo Vilalta,et al.  Introduction to the Special Issue on Meta-Learning , 2004, Machine Learning.

[2]  David H. Wolpert,et al.  Stacked generalization , 1992, Neural Networks.

[3]  Bogdan Gabrys,et al.  Application of Computational Intelligence Techniques to Process Industry Problems , 2008 .

[4]  Thomas G. Dietterich Multiple Classifier Systems , 2000, Lecture Notes in Computer Science.

[5]  Ricardo Vilalta,et al.  A Perspective View and Survey of Meta-Learning , 2002, Artificial Intelligence Review.

[6]  Yves Chauvin,et al.  Back-Propagation: Theory, Architecture, and Applications , 1995 .

[7]  Subhash C. Bagui,et al.  Combining Pattern Classifiers: Methods and Algorithms , 2005, Technometrics.

[8]  Alan F. Murray,et al.  International Joint Conference on Neural Networks , 1993 .

[9]  John Hallam,et al.  IEEE International Joint Conference on Neural Networks , 2005 .

[10]  Sargur N. Srihari,et al.  Decision Combination in Multiple Classifier Systems , 1994, IEEE Trans. Pattern Anal. Mach. Intell..

[11]  Michael I. Jordan,et al.  Task Decomposition Through Competition in a Modular Connectionist Architecture: The What and Where Vision Tasks , 1990, Cogn. Sci..

[12]  Luigi Fortuna,et al.  Soft Sensors for Monitoring and Control of Industrial Processes (Advances in Industrial Control) , 2006 .

[13]  Andrew W. Moore,et al.  Locally Weighted Learning , 1997, Artificial Intelligence Review.

[14]  Geoffrey E. Hinton,et al.  Adaptive Mixtures of Local Experts , 1991, Neural Computation.

[15]  Christophe Giraud-Carrier Beyond predictive accuracy : what? , 1998 .

[16]  Jiri Matas,et al.  On Combining Classifiers , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[17]  Christophe G. Giraud-Carrier,et al.  Predicting Algorithm Accuracy with a Small Set of Effective Meta-Features , 2008, 2008 Seventh International Conference on Machine Learning and Applications.

[18]  Michael I. Jordan,et al.  Task Decompostiion Through Competition in a Modular Connectionist Architecture: The What and Where Vision Tasks , 1993, Machine Learning: From Theory to Applications.

[19]  D. Ruta,et al.  An Overview of Classifier Fusion Methods , 2000 .

[20]  Heekuck Oh,et al.  Neural Networks for Pattern Recognition , 1993, Adv. Comput..

[21]  Christopher G. Atkeson,et al.  Constructive Incremental Learning from Only Local Information , 1998, Neural Computation.

[22]  David W. Aha,et al.  Generalizing from Case studies: A Case Study , 1992, ML.

[23]  Bernhard Pfahringer,et al.  Locally Weighted Naive Bayes , 2002, UAI.