Random Prism: An Alternative to Random Forests

Ensemble learning techniques generate multiple classifiers, so called base classifiers, whose combined classification results are used in order to increase the overall classification accuracy. In most ensemble classifiers the base classifiers are based on the Top Down Induction of Decision Trees (TDIDT) approach. However, an alternative approach for the induction of rule based classifiers is the Prism family of algorithms. Prism algorithms produce modular classification rules that do not necessarily fit into a decision tree structure. Prism classification rulesets achieve a comparable and sometimes higher classification accuracy compared with decision tree classifiers, if the data is noisy and large. Yet Prism still suffers from overfitting on noisy and large datasets. In practice ensemble techniques tend to reduce the overfitting, however there exists no ensemble learner for modular classification rule inducers such as the Prism family of algorithms. This article describes the first development of an ensemble learner based on the Prism family of algorithms in order to enhance Prism’s classification accuracy by reducing overfitting.

[1]  Max Bramer,et al.  Inducer: a public domain workbench for data mining , 2005, Int. J. Syst. Sci..

[2]  Mo Adda,et al.  Parallel Induction of Modular Classification Rules , 2008, SGAI Conf..

[3]  Philip S. Yu,et al.  Distributed hoeffding trees for pocket data mining , 2011, 2011 International Conference on High Performance Computing & Simulation.

[4]  Padhraic Smyth,et al.  An Information Theoretic Approach to Rule Induction from Databases , 1992, IEEE Trans. Knowl. Data Eng..

[5]  Philip S. Yu,et al.  Distributed Classification for Pocket Data Mining , 2011, ISMIS.

[6]  Max Bramer,et al.  Induction of Modular Classification Rules: Using Jmax-pruning , 2010, SGAI Conf..

[7]  Frans Coenen,et al.  Research and Development in Intelligent Systems XVI , 2000, Springer London.

[8]  Geoff Hulten,et al.  Mining high-speed data streams , 2000, KDD '00.

[9]  Salvatore J. Stolfo,et al.  Experiments on multistrategy learning by meta-learning , 1993, CIKM '93.

[10]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[11]  Bernard Zenko,et al.  Is Combining Classifiers with Stacking Better than Selecting the Best One? , 2004, Machine Learning.

[12]  Max Bramer,et al.  An Information-Theoretic Approach to the Pre-pruning of Classification Rules , 2002, Intelligent Information Processing.

[13]  Ian H. Witten,et al.  Data mining: practical machine learning tools and techniques, 3rd Edition , 1999 .

[14]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[15]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[16]  Foster Provost,et al.  Distributed Data Mining: Scaling up and beyond , 2000 .

[17]  Philip S. Yu,et al.  Pocket Data Mining: Towards Collaborative Data Mining in Mobile Computing Environments , 2010, 2010 22nd IEEE International Conference on Tools with Artificial Intelligence.

[18]  Roberto J. Bayardo,et al.  PLANET: Massively Parallel Learning of Tree Ensembles with MapReduce , 2009, Proc. VLDB Endow..

[19]  Mo Adda,et al.  J-PMCRI: A Methodology for Inducing Pre-pruned Modular Classification Rules , 2010, IFIP AI.

[20]  Bernard Zenko,et al.  Is Combining Classifiers Better than Selecting the Best One , 2002, ICML.

[21]  Nitesh V. Chawla,et al.  Learning Ensembles from Bites: A Scalable and Accurate Approach , 2004, J. Mach. Learn. Res..

[22]  Max Bramer,et al.  Automatic Induction of Classification Rules from Examples Using N-Prism , 2000 .

[23]  J. Ross Quinlan,et al.  C4.5: Programs for Machine Learning , 1992 .

[24]  Mo Adda,et al.  PMCRI: A Parallel Modular Classification Rule Induction Framework , 2009, MLDM.

[25]  Ryszard S. Michalski,et al.  On the Quasi-Minimal Solution of the General Covering Problem , 1969 .

[26]  Ian H. Witten,et al.  Data mining: practical machine learning tools and techniques with Java implementations , 2002, SGMD.

[27]  Jadzia Cendrowska,et al.  PRISM: An Algorithm for Inducing Modular Rules , 1987, Int. J. Man Mach. Stud..

[28]  金田 重郎,et al.  C4.5: Programs for Machine Learning (書評) , 1995 .

[29]  Sanjay Ghemawat,et al.  MapReduce: Simplified Data Processing on Large Clusters , 2004, OSDI.

[30]  Mo Adda,et al.  Parallel Rule Induction with Information Theoretic Pre-Pruning , 2009, SGAI Conf..

[31]  J. Ross Quinlan,et al.  Induction of Decision Trees , 1986, Machine Learning.

[32]  Jiawei Han,et al.  Data Mining: Concepts and Techniques , 2000 .