The data mining advisor: meta-learning at the service of practitioners

In order to make machine learning algorithms more usable, our community must be able to design robust systems that offer support to practitioners. In the context of classification, this amounts to developing assistants, which deal with the increasing number of models and techniques, and give advice dynamically on such issues as model selection and method combination. This paper briefly reviews the potential of meta-learning in this context and reports on the early success of a Web-based data mining assistant.

[1]  So Young Sohn,et al.  Meta Analysis of Classification Algorithms for Pattern Recognition , 1999, IEEE Trans. Pattern Anal. Mach. Intell..

[2]  Peter A. Flach,et al.  Improved Dataset Characterisation for Meta-learning , 2002, Discovery Science.

[3]  Robert C. Holte,et al.  Very Simple Classification Rules Perform Well on Most Commonly Used Datasets , 1993, Machine Learning.

[4]  Hilan Bensusan,et al.  A Higher-order Approach to Meta-learning , 2000, ILP Work-in-progress reports.

[5]  D. Wolpert,et al.  No Free Lunch Theorems for Search , 1995 .

[6]  Maarten van Someren Model Class Selection and Construction: Beyond the Procrustean Approach to Machine Learning Applications , 2001, Machine Learning and Its Applications.

[7]  Hilan Bensusan,et al.  Meta-Learning by Landmarking Various Learning Algorithms , 2000, ICML.

[8]  Hilan Bensusan Odd bites into bananas don''t make you blind: learning about simplicity and attribute addition , 1998 .

[9]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[10]  Melanie Hilario,et al.  Model selection via meta-learning: a comparative study , 2000, Proceedings 12th IEEE Internationals Conference on Tools with Artificial Intelligence. ICTAI 2000.

[11]  Hilan Bensusan,et al.  Estimating the Predictive Accuracy of a Classifier , 2001, ECML.

[12]  Iain Paterson,et al.  Evaluation of Machine-Learning Algorithm Ranking Advisors , 2000 .

[13]  Aiko M. Hormann,et al.  Programs for Machine Learning. Part I , 1962, Inf. Control..

[14]  Carlos Soares,et al.  A Comparison of Ranking Methods for Classification Algorithm Selection , 2000, ECML.

[15]  Rudi Studer,et al.  AST: Support for Algorithm Selection with a CBR Approach , 1999, PKDD.

[16]  P. Andersen,et al.  A procedure for ranking efficient units in data envelopment analysis , 1993 .

[17]  Robert Engels,et al.  Using a Data Metric for Preprocessing Advice for Data Mining Applications , 1998, ECAI.

[18]  David J. Spiegelhalter,et al.  Machine Learning, Neural and Statistical Classification , 2009 .

[19]  Melanie Hilario,et al.  Fusion of Meta-knowledge and Meta-data for Case-Based Model Selection , 2001, PKDD.

[20]  David H. Wolpert Any Two Learning Algorithms Are (Almost) Exactly Identical , 2000 .

[21]  Hilan Bensusan,et al.  Discovering Task Neighbourhoods Through Landmark Learning Performances , 2000, PKDD.

[22]  Ricardo Vilalta,et al.  Using Meta-Learning to Support Data Mining , 2004, Int. J. Comput. Sci. Appl..

[23]  Cullen Schaffer,et al.  A Conservation Law for Generalization Performance , 1994, ICML.

[24]  Johannes Fürnkranz,et al.  An Evaluation of Landmarking Variants , 2001 .

[25]  D. Wolpert The Supervised Learning No-Free-Lunch Theorems , 2002 .

[26]  Carlos Soares,et al.  Ranking Learning Algorithms: Using IBL and Meta-Learning on Accuracy and Time Results , 2003, Machine Learning.

[27]  Eamonn J. Keogh,et al.  Towards parameter-free data mining , 2004, KDD.

[28]  Christophe Giraud-Carrier Beyond predictive accuracy : what? , 1998 .

[29]  J. Ross Quinlan,et al.  C4.5: Programs for Machine Learning , 1992 .

[30]  Ian H. Witten,et al.  Data mining: practical machine learning tools and techniques with Java implementations , 2002, SGMD.

[31]  Eamonn J. Keogh,et al.  UCR Time Series Data Mining Archive , 1983 .