Very Fast Decision Rules for multi-class problems

Decision rules are one of the most interpretable and flexible models for data mining prediction tasks. Till now, few works presented online, any-time and one-pass algorithms for learning decision rules in the stream mining scenario. A quite recent algorithm, the Very Fast Decision Rules (VFDR), learns set of rules, where each rule discriminates one class from all the other. In this work we extend the VFDR algorithm by decomposing a multi-class problem into a set of two-class problems and inducing a set of discriminative rules for each binary problem. The proposed algorithm maintains all properties required when learning from stationary data streams: online and any-time classifiers, processing each example once. Moreover, it is able to learn ordered and unordered rule sets. The new approach is evaluated on various real and artificial datasets. The new algorithm improves the performance of the previous version and is competitive with the state-of-the-art decision tree learning method for data streams.

[1]  Gerhard Widmer,et al.  Learning in the Presence of Concept Drift and Hidden Contexts , 1996, Machine Learning.

[2]  GamaJoão,et al.  Decision trees for mining data streams , 2006 .

[3]  William W. Cohen Fast Effective Rule Induction , 1995, ICML.

[4]  Frank Klawonn,et al.  Guide to Intelligent Data Analysis - How to Intelligently Make Sense of Real Data , 2010, Texts in Computer Science.

[5]  Geoff Hulten,et al.  Mining time-changing data streams , 2001, KDD '01.

[6]  J. C. Schlimmer,et al.  Incremental learning from noisy data , 2004, Machine Learning.

[7]  João Gama,et al.  Learning Decision Rules from Data Streams , 2011, IJCAI.

[8]  Albert Bifet,et al.  Massive Online Analysis , 2009 .

[9]  Myeong-Kwan Kevin Cheon,et al.  Frank and I , 2012 .

[10]  Pedro M. Domingos Unifying Instance-Based and Rule-Based Induction , 1996, Machine Learning.

[11]  Johannes Fürnkranz,et al.  Round Robin Rule Learning , 2001, ICML.

[12]  Ryszard S. Michalski,et al.  Incremental learning with partial instance memory , 2002, Artif. Intell..

[13]  Jesús S. Aguilar-Ruiz,et al.  Incremental Rule Learning and Border Examples Selection from Numerical Data Streams , 2005, J. Univers. Comput. Sci..

[14]  Peter Clark,et al.  Rule Induction with CN2: Some Recent Improvements , 1991, EWSL.

[15]  J. Ross Quinlan,et al.  C4.5: Programs for Machine Learning , 1992 .

[16]  Peter Clark,et al.  The CN2 Induction Algorithm , 1989, Machine Learning.

[17]  João Gama,et al.  Accurate decision trees for mining high-speed data streams , 2003, KDD '03.

[18]  Ian H. Witten,et al.  Generating Accurate Rule Sets Without Global Optimization , 1998, ICML.

[19]  Alberto Maria Segre,et al.  Programs for Machine Learning , 1994 .

[20]  Geoff Hulten,et al.  Mining high-speed data streams , 2000, KDD '00.

[21]  Marcus A. Maloof,et al.  Dynamic weighted majority: a new ensemble method for tracking concept drift , 2003, Third IEEE International Conference on Data Mining.

[22]  Ronald L. Rivest,et al.  Learning decision lists , 2004, Machine Learning.

[23]  Sholom M. Weiss,et al.  Predictive data mining - a practical guide , 1997 .

[24]  João Gama,et al.  Decision trees for mining data streams , 2006, Intell. Data Anal..

[25]  William Nick Street,et al.  A streaming ensemble algorithm (SEA) for large-scale classification , 2001, KDD '01.

[26]  Geoff Holmes,et al.  MOA: Massive Online Analysis , 2010, J. Mach. Learn. Res..

[27]  R. Mike Cameron-Jones,et al.  Induction of logic programs: FOIL and related systems , 1995, New Generation Computing.