MofN rule extraction from neural networks trained with augmented discretized input

The accuracy of neural networks can be improved when they are trained with discretized continuous attributes as additional inputs. Such input augmentation makes it easier for the network weights to form more accurate decision boundaries when the data samples of different classes in the data set are contained in distinct hyper-rectangular subregions in the original input space. In this paper we present first how a neural network can be trained with augmented discretized inputs. The additional inputs are obtained by simply dividing the original interval of each continuous attribute into subintervals of equal length. Thermometer encoding scheme is used to represent these discretized inputs. The network is then pruned to remove most of the discretized inputs as well as the original continuous attributes as long as the network still achieves a minimum preset accuracy requirement. We then discuss how MofN rules can be extracted from the pruned network by analyzing the activations of the network's hidden units and the weights of the network connections that remain in the pruned network. For data sets that have sample classes defined by relatively complex boundaries, surprisingly simple MofN rules with very good accuracy rates are obtained.

[1]  Engelbert Mephu Nguifo,et al.  Towards a generalization of decompositional approach of rule extraction from multilayer artificial neural network , 2011, The 2011 International Joint Conference on Neural Networks.

[2]  Huan Liu,et al.  Chi2: feature selection and discretization of numeric attributes , 1995, Proceedings of 7th IEEE International Conference on Tools with Artificial Intelligence.

[3]  Richi Nayak Generating rules with predicates, terms and variables from the pruned neural networks , 2009, Neural Networks.

[4]  Jacek M. Zurada,et al.  Extracting Rules From Neural Networks as Decision Diagrams , 2011, IEEE Transactions on Neural Networks.

[5]  Murray Smith,et al.  Neural Networks for Statistical Modeling , 1993 .

[6]  T. Kathirvalavakumar,et al.  Rule extraction from neural networks — A comparative study , 2012, International Conference on Pattern Recognition, Informatics and Medical Engineering (PRIME-2012).

[7]  Emilio Carrizosa,et al.  Detecting relevant variables and interactions in supervised classification , 2011, Eur. J. Oper. Res..

[8]  Hari Om Gupta,et al.  Function analysis based rule extraction from artificial neural networks for transformer incipient fault diagnosis , 2012 .

[9]  A. E. Amin,et al.  A novel classification model for cotton yarn quality based on trained neural network using genetic algorithm , 2013, Knowl. Based Syst..

[10]  Aiko M. Hormann,et al.  Programs for Machine Learning. Part I , 1962, Inf. Control..

[11]  Vincent A. Schmidt,et al.  Connectionist-Based Rules Describing the Pass-Through of Individual Goods Prices into Trend Inflation in the United States , 2010, IC-AI.

[12]  David McLean,et al.  Rule extraction from neural networks for medical domains , 2010, The 2010 International Joint Conference on Neural Networks (IJCNN).

[13]  Jude W. Shavlik,et al.  Extracting Refined Rules from Knowledge-Based Neural Networks , 1993, Machine Learning.

[14]  Chee Peng Lim,et al.  Improved GART Neural Network Model for Pattern Classification and Rule Extraction With Application to Power Systems , 2011, IEEE Transactions on Neural Networks.

[15]  D. K. Mishra,et al.  KDRuleEx: A Novel Approach for Enhancing User Comprehensibility Using Rule Extraction , 2012, 2012 Third International Conference on Intelligent Systems Modelling and Simulation.

[16]  Mohammad Saniee Abadeh,et al.  Fuzzy Rule Extraction from a trained artificial neural network using Genetic Algorithm for WECS control and parameter estimation , 2011, 2011 Eighth International Conference on Fuzzy Systems and Knowledge Discovery (FSKD).

[17]  Liping Su A Rule Extraction Based Approach in Predicting Derivative Use for Financial Risk Hedging in Construction Companies , 2011 .

[18]  Joachim Diederich,et al.  The truth will come to light: directions and challenges in extracting the knowledge embedded within trained artificial neural networks , 1998, IEEE Trans. Neural Networks.

[19]  Rudy Setiono Extracting M-of-N rules from trained neural networks , 2000, IEEE Trans. Neural Networks Learn. Syst..

[20]  Jacek M. Zurada,et al.  White box nonlinear prediction models, editorial special issue , 2011 .

[21]  Marko Robnik-Sikonja,et al.  Explaining Classifications For Individual Instances , 2008, IEEE Transactions on Knowledge and Data Engineering.

[22]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[23]  Rudy Setiono,et al.  A Penalty-Function Approach for Pruning Feedforward Neural Networks , 1997, Neural Computation.

[24]  Huan Liu,et al.  Symbolic Representation of Neural Networks , 1996, Computer.

[25]  Rudy Setiono,et al.  Discrete Variable Generation for Improved Neural Network Classification , 2012, 2012 IEEE 24th International Conference on Tools with Artificial Intelligence.

[26]  Kit Yan Chan,et al.  Identification of significant factors for air pollution levels using a neural network based knowledge discovery system , 2013, Neurocomputing.

[27]  Huan Liu,et al.  A connectionist approach to generating oblique decision trees , 1999, IEEE Trans. Syst. Man Cybern. Part B.

[28]  So-Jung Park,et al.  Rule Extraction for Dynamic Hand Gesture Recognition using a Modified FMM Neural Network , 2013 .