Recursive Automatic Bias Selection for Classifier Construction

The results of empirical comparisons of existing learning algorithms illustrate that each algorithm has a selective superiority; each is best for some but not all tasks. Given a data set, it is often not clear beforehand which algorithm will yield the best performance. In this article we present an approach that uses characteristics of the given data set, in the form of feedback from the learning process, to guide a search for a tree-structured hybrid classifier. Heuristic knowledge about the characteristics that indicate one bias is better than another is encoded in the rule base of the Model Class Selection (MCS) system. The approach does not assume that the entire instance space is best learned using a single representation language; for some data sets, choosing to form a hybrid classifier is a better bias, and MCS has the ability to determine these cases. The results of an empirical evaluation illustrate that MCS achieves classification accuracies equal to or higher than the best of its primitive learning components for each data set, demonstrating that the heuristic rules effectively select an appropriate learning bias.

[1]  Richard O. Duda,et al.  Pattern classification and scene analysis , 1974, A Wiley-Interscience publication.

[2]  J. Ross Quinlan,et al.  Simplifying Decision Trees , 1987, Int. J. Man Mach. Stud..

[3]  Paul E. Utgoff,et al.  Perceptron Trees : A Case Study in ybrid Concept epresentations , 1999 .

[4]  Sholom M. Weiss,et al.  An Empirical Comparison of Pattern Recognition, Neural Nets, and Machine Learning Classification Methods , 1989, IJCAI.

[5]  R. Detrano,et al.  International application of a new probability algorithm for the diagnosis of coronary artery disease. , 1989, The American journal of cardiology.

[6]  Larry A. Rendell,et al.  Building Robust Learning Systems by Combining Induction and Optimization , 1989, IJCAI.

[7]  David W. Aha,et al.  A study of instance-based algorithms for supervised learning tasks: mathematical, empirical, and psychological evaluations , 1990 .

[8]  Marcus R. Frean,et al.  Small nets and short paths : optimising neural computation , 1990 .

[9]  Carla E. Brodley,et al.  Linear Machine Decision Trees , 1991 .

[10]  Foster J. Provost,et al.  Inductive Policy , 1992, AAAI.

[11]  J. Mesirov,et al.  Hybrid system for protein secondary structure prediction. , 1992, Journal of molecular biology.

[12]  David W. Aha,et al.  Generalizing from Case studies: A Case Study , 1992, ML.

[13]  David H. Wolpert,et al.  Stacked generalization , 1992, Neural Networks.

[14]  Stephen C. Y. Lu,et al.  Creating and using models for engineering design: a machine-learning approach , 1992, IEEE Expert.

[15]  J. Ross Quinlan,et al.  Combining Instance-Based and Model-Based Learning , 1993, ICML.

[16]  Carla E. Brodley,et al.  Addressing the Selective Superiority Problem: Automatic Algorithm/Model Class Selection , 1993 .

[17]  Carla E. Brodley Recursive automatic algorithm selection for inductive learning , 1995 .

[18]  Jorma Rissanen,et al.  Stochastic Complexity in Statistical Inquiry , 1989, World Scientific Series in Computer Science.

[19]  Cullen Schaffer,et al.  Selecting a classification method by cross-validation , 1993, Machine Learning.

[20]  Larry A. Rendell,et al.  Empirical learning as a function of concept character , 2004, Machine Learning.

[21]  A. Meyer-Bäse Feature Selection and Extraction , 2004 .

[22]  Carla E. Brodley,et al.  Multivariate decision trees , 2004, Machine Learning.

[23]  J. Ross Quinlan,et al.  Induction of Decision Trees , 1986, Machine Learning.

[24]  Steven Salzberg,et al.  A Nearest Hyperrectangle Learning Method , 1991, Machine Learning.

[25]  David W. Aha,et al.  Instance-Based Learning Algorithms , 1991, Machine Learning.