Separate-and-conquer Regression Technical Report TUD – KE – 2010 – 01

In this paper a rule learning algorithm for the prediction of numerical target variables is presented. It is based on the separate-and-conquer strategy and the classification phase is done by a decision list. A new splitpoint generation method is introduced for the efficient handling of numerical attributes. It is shown that the algorithm performs comparable to other regression algorithms where some of them are based on rules and some are not. Additionally a novel heuristic for evaluating the trade-off between consistency and generality of regression rules is introduced. This heuristic features a parameter to directly trade off the rules consistency and its generality. We present an optimal setting for this parameter based on optimizing it on several data sets. The algorithm features two additional parameters that are also tuned on the same datasets as the heuristics parameter. The evaluation part of the paper gives insights on results obtained on tuning datasets that were split into two folds of equal size. The algorithm was tuned on the first set of these split databases and is evaluated on the hold-out folds and vice versa yielding two configurations of the rule learner. These are also evaluated on 9 testing datasets that were not used during the optimization.

[1]  Bogdan E. Popescu,et al.  PREDICTIVE LEARNING VIA RULE ENSEMBLES , 2008, 0811.1679.

[2]  Johannes Fürnkranz,et al.  Incremental Reduced Error Pruning , 1994, ICML.

[3]  R. Mike Cameron-Jones,et al.  Oversearching and Layered Search in Empirical Learning , 1995, IJCAI.

[4]  Peter Clark,et al.  The CN2 Induction Algorithm , 1989, Machine Learning.

[5]  JOHANNES FÜRNKRANZ,et al.  Separate-and-Conquer Rule Learning , 1999, Artificial Intelligence Review.

[6]  Johannes Fürnkranz,et al.  An Empirical Investigation of the Trade-Off between Consistency and Coverage in Rule Learning Heuristics , 2008, Discovery Science.

[7]  Janez Demsar,et al.  Statistical Comparisons of Classifiers over Multiple Data Sets , 2006, J. Mach. Learn. Res..

[8]  Luís Torgo,et al.  DATA FITTING WITH RULE-BASED REGRESSION , 2007 .

[9]  J. R. Quinlan Learning With Continuous Classes , 1992 .

[10]  Johannes Fürnkranz,et al.  ROC ‘n’ Rule Learning—Towards a Better Understanding of Covering Algorithms , 2005, Machine Learning.

[11]  Peter A. Flach,et al.  Subgroup Discovery with CN2-SD , 2004, J. Mach. Learn. Res..

[12]  Geoff Holmes,et al.  Generating Rule Sets from Model Trees , 1999, Australian Joint Conference on Artificial Intelligence.

[13]  Johannes Fürnkranz,et al.  A Re-evaluation of the Over-Searching Phenomenon in Inductive Rule Learning , 2008, LWA.

[14]  Wojciech Kotlowski,et al.  Solving Regression by Learning an Ensemble of Decision Rules , 2006, ICAISC.

[15]  Ian H. Witten,et al.  Induction of model trees for predicting continuous classes , 1996 .

[16]  Steven Salzberg,et al.  Lookahead and Pathology in Decision Tree Induction , 1995, IJCAI.

[17]  Sholom M. Weiss,et al.  Rule-based Machine Learning Methods for Functional Prediction , 1995, J. Artif. Intell. Res..

[18]  Luís Torgo,et al.  Regression by Classification , 1996, SBIA.

[19]  Peter Clark,et al.  Rule Induction with CN2: Some Recent Improvements , 1991, EWSL.

[20]  Ian H. Witten,et al.  Data mining: practical machine learning tools and techniques with Java implementations , 2002, SGMD.

[21]  Ivan Bratko,et al.  First Order Regression , 1997, Machine Learning.

[22]  Robert C. Holte,et al.  Concept Learning and the Problem of Small Disjuncts , 1989, IJCAI.

[23]  William W. Cohen Fast Effective Rule Induction , 1995, ICML.

[24]  Bernard Ženko,et al.  Learning Predictive Clustering Rules , 2005, Informatica.