Dynamic Class Prediction with Classifier Based Distance Measure

Combining multiple classifiers (ensemble of classifiers) to make predictions for new instances has shown to outperform a single classifier. As opposed to using the same ensemble for all data instances, recent studies have focused on dynamic ensembles in which a new ensemble is chosen from a pool of classifiers specifically for every new data instance. We propose a system for dynamic class prediction based on a new distance measure to evaluate the distance among data instances. We first map data instances into a space defined by the class probability estimates from a pool of two-class classifiers. We dynamically pick classifiers (features) to be used and the k-nearest neighbors of a new instance by minimizing the distance between the neighbors and that instance in a two-step framework. Results of our experiments show that our measure is effective for finding similar instances and our framework helps making more accurate predictions.

[1]  William Nick Street,et al.  Ensemble Pruning Via Semi-definite Programming , 2006, J. Mach. Learn. Res..

[2]  Luca Didaci,et al.  Dynamic Classifier Selection by Adaptive k-Nearest-Neighbourhood Rule , 2004, Multiple Classifier Systems.

[3]  Ian H. Witten,et al.  Stacking Bagged and Dagged Models , 1997, ICML.

[4]  Robert Sabourin,et al.  Dynamic Selection of Ensembles of Classifiers Using Contextual Information , 2010, MCS.

[5]  Luiz S. Oliveira,et al.  Improving a dynamic ensemble selection method based on oracle information , 2012 .

[6]  Fabio Roli,et al.  Dynamic classifier selection based on multiple classifier behaviour , 2001, Pattern Recognit..

[7]  David H. Wolpert,et al.  Stacked generalization , 1992, Neural Networks.

[8]  Gian Luca Marcialis,et al.  A study on the performances of dynamic classifier selection based on local accuracy estimation , 2005, Pattern Recognit..

[9]  Chih-Jen Lin,et al.  LIBSVM: A library for support vector machines , 2011, TIST.

[10]  Kagan Tumer,et al.  Error Correlation and Error Reduction in Ensemble Classifiers , 1996, Connect. Sci..

[11]  Thomas G. Dietterich What is machine learning? , 2020, Archives of Disease in Childhood.

[12]  Yoav Freund,et al.  Experiments with a New Boosting Algorithm , 1996, ICML.

[13]  Robert Sabourin,et al.  From dynamic classifier selection to dynamic ensemble selection , 2008, Pattern Recognit..

[14]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[15]  Lior Rokach,et al.  Troika - An improved stacking schema for classification tasks , 2009, Inf. Sci..

[16]  Kevin W. Bowyer,et al.  Combination of multiple classifiers using local accuracy estimates , 1996, Proceedings CVPR IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[17]  Robert Sabourin,et al.  A dynamic overproduce-and-choose strategy for the selection of classifier ensembles , 2008, Pattern Recognit..

[18]  Tin Kam Ho,et al.  The Random Subspace Method for Constructing Decision Forests , 1998, IEEE Trans. Pattern Anal. Mach. Intell..