Regression rules extraction from artificial neural network based on least squares

As an intelligent approach, the artificial neural network (ANN) has been shown to exhibit superior predictive power compared to traditional approaches in many studies. However, its further development is restricted by its characteristic of “black box” because it can't explain how learning from input data was done and how performance can be consistently ensured. In order to open the “black box”, a novel approach of regression rules extraction from ANN based on Least Squares is proposed in this paper, it can uncover underlying dependencies between the input data and the output data of ANN. Its main concepts are that the nonlinear activation function of each hidden unit is approximated locally by a three-piece linear function based on Least Squares and then the regression rules are generated by decision tree approach. The piecewise linear regression rules can not only ensure the accuracy but also enhance the explanation. Experiments on two public datasets show that the proposed approach generates rules are more accurate than the existing approaches based on decision trees or linear regression.

[1]  Kwan-Liu Ma,et al.  Opening the black box - data driven visualization of neural networks , 2005, VIS 05. IEEE Visualization, 2005..

[2]  Rudy Setiono,et al.  Extracting Rules from Neural Networks by Pruning and Hidden-Unit Splitting , 1997, Neural Computation.

[3]  M. Stone Cross‐Validatory Choice and Assessment of Statistical Predictions , 1976 .

[4]  Alberto Maria Segre,et al.  Programs for Machine Learning , 1994 .

[5]  Mohammad Bagher Menhaj,et al.  Training feedforward networks with the Marquardt algorithm , 1994, IEEE Trans. Neural Networks.

[6]  Wang Jian-guo,et al.  Rule extraction from artificial neural network with optimized activation functions , 2008, 2008 3rd International Conference on Intelligent System and Knowledge Engineering.

[7]  Jacek M. Zurada,et al.  Extraction of rules from artificial neural networks for nonlinear regression , 2002, IEEE Trans. Neural Networks.

[8]  Julian D. Olden,et al.  Illuminating the “black box”: a randomization approach for understanding variable contributions in artificial neural networks , 2002 .

[9]  B. Schenker,et al.  Cross-validated structure selection for neural networks , 1996 .

[10]  Tim Andersen,et al.  Cross validation and MLP architecture selection , 1999, IJCNN'99. International Joint Conference on Neural Networks. Proceedings (Cat. No.99CH36339).

[11]  LiMin Fu,et al.  Rule Generation from Neural Networks , 1994, IEEE Trans. Syst. Man Cybern. Syst..

[12]  Paul-Henry Cournède,et al.  Structural identifiability of generalized constraint neural network models for nonlinear regression , 2008, Neurocomputing.

[13]  Jude W. Shavlik,et al.  Knowledge-Based Artificial Neural Networks , 1994, Artif. Intell..

[14]  Rudy Setiono,et al.  An Approach To Generate Rules From Neural Networks for Regression Problems , 2004, Eur. J. Oper. Res..

[15]  Joachim Diederich,et al.  Survey and critique of techniques for extracting rules from trained artificial neural networks , 1995, Knowl. Based Syst..