Induction of Modular Classification Rules by Information Entropy Based Rule Generation

Prism has been developed as a modular classification rule generator following the separate and conquer approach since 1987 due to the replicated sub-tree problem occurring in Top-Down Induction of Decision Trees (TDIDT). A series of experiments have been done to compare the performance between Prism and TDIDT which proved that Prism may generally provide a similar level of accuracy as TDIDT but with fewer rules and fewer terms per rule. In addition, Prism is generally more tolerant to noise with consistently better accuracy than TDIDT. However, the authors have identified through some experiments that Prism may also give rule sets which tend to underfit training sets in some cases. This paper introduces a new modular classification rule generator, which follows the separate and conquer approach, in order to avoid the problems which arise with Prism. In this paper, the authors review the Prism method and its advantages compared with TDIDT as well as its disadvantages that are overcome by a new method using Information Entropy Based Rule Generation (IEBRG). The authors also set up an experimental study on the performance of the new method in classification accuracy and computational efficiency. The method is also evaluated comparatively with Prism.

[1]  Max Bramer,et al.  Computationally efficient induction of classification rules with the PMCRI and J-PMCRI frameworks , 2012, Knowl. Based Syst..

[2]  Max Bramer,et al.  Induction of Modular Classification Rules: Using Jmax-pruning , 2010, SGAI Conf..

[3]  Aiko M. Hormann,et al.  Programs for Machine Learning. Part I , 1962, Inf. Control..

[4]  Philip J. Stone,et al.  Experiments in induction , 1966 .

[5]  Max Bramer,et al.  Inducer: a public domain workbench for data mining , 2005, Int. J. Syst. Sci..

[6]  Max Bramer,et al.  Using J-pruning to reduce overfitting in classification trees , 2002, Knowl. Based Syst..

[7]  J. Ross Quinlan,et al.  C4.5: Programs for Machine Learning , 1992 .

[8]  Ryszard S. Michalski,et al.  On the Quasi-Minimal Solution of the General Covering Problem , 1969 .

[9]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[10]  Randy Kerber,et al.  ChiMerge: Discretization of Numeric Attributes , 1992, AAAI.

[11]  Padhraic Smyth,et al.  9 Rule Induction using Information Theory , 2002 .

[12]  Jadzia Cendrowska,et al.  PRISM: An Algorithm for Inducing Modular Rules , 1987, Int. J. Man Mach. Stud..

[13]  Max Bramer,et al.  Jmax-pruning: A facility for the information theoretic pruning of modular classification rules , 2012, Knowl. Based Syst..

[14]  Max Bramer,et al.  Principles of Data Mining , 2016, Undergraduate Topics in Computer Science.

[15]  Max Bramer,et al.  Automatic Induction of Classification Rules from Examples Using N-Prism , 2000 .

[16]  C. E. SHANNON,et al.  A mathematical theory of communication , 1948, MOCO.

[17]  Max Bramer,et al.  Using J-Pruning to Reduce Overfitting of Classification Rules in Noisy Domains , 2002, DEXA.