A novel ensemble method for k-nearest neighbor

Abstract In this paper, to address the issue that ensembling k-nearest neighbor (kNN) classifiers with resampling approaches cannot generate component classifiers with a large diversity, we consider ensembling kNN through a multimodal perturbation-based method. Since kNN is sensitive to the input attributes, we propose a weighted heterogeneous distance Metric (WHDM). By using a WHDM and evidence theory, a progressive kNN classifier is developed. Based on a progressive kNN, the random subspace method, attribute reduction, and Bagging, a novel algorithm termed RRSB (reduced random subspace-based Bagging) is proposed for construct ensemble classifier, which can increase the diversity of component classifiers without damaging the accuracy of the component classifiers. In detail, RRSB adopts the perturbation on the learning parameter with a weighted heterogeneous distance metric, the perturbation on the input space with random subspace and attribute reduction, the perturbation on the training data with Bagging, and the perturbation on the output target of k neighbors with evidence theory. In the experimental stage, the value of k, the different perturbations on RRSB and the ensemble size are analyzed. In addition, RRSB is compared with other multimodal perturbation-based ensemble algorithms on multiple UCI data sets and a KDD data set. The results from the experiments demonstrate the effectiveness of RRSB for kNN ensembling.

[1]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[2]  Guoqiang Peter Zhang,et al.  Neural networks for classification: a survey , 2000, IEEE Trans. Syst. Man Cybern. Part C.

[3]  Minqiang Li,et al.  Coevolutionary learning of neural network ensemble for complex classification tasks , 2012, Pattern Recognit..

[4]  Naohiro Ishii,et al.  Combining classification improvements by ensemble processing , 2005, Third ACIS Int'l Conference on Software Engineering Research, Management and Applications (SERA'05).

[5]  Qinghua Hu,et al.  Efficient multi-modal geometric mean metric learning , 2018, Pattern Recognit..

[6]  Lin Feng,et al.  Multi-view metric learning based on KL-divergence for similarity measurement , 2017, Neurocomputing.

[7]  D. Opitz,et al.  Popular Ensemble Methods: An Empirical Study , 1999, J. Artif. Intell. Res..

[8]  Wen Gao,et al.  Parametric local multiview hamming distance metric learning , 2018, Pattern Recognit..

[9]  Tiranee Achalakul,et al.  Reducing bioinformatics data dimension with ABC-kNN , 2013, Neurocomputing.

[10]  Hakan Altinçay,et al.  Ensembling evidential k-nearest neighbor classifiers through multi-modal perturbation , 2007, Appl. Soft Comput..

[11]  Thierry Denoeux,et al.  A k-nearest neighbor classification rule based on Dempster-Shafer theory , 1995, IEEE Trans. Syst. Man Cybern..

[12]  Jerzy W. Grzymala-Busse,et al.  Rough Sets , 1995, Commun. ACM.

[13]  Juan Ramón Rico-Juan,et al.  Improving kNN multi-label classification in Prototype Selection scenarios using class proposals , 2015, Pattern Recognit..

[14]  Michael K. Ng,et al.  Automated variable weighting in k-means type clustering , 2005, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[15]  Thomas G. Dietterich Machine-Learning Research , 1997, AI Mag..

[16]  Tony R. Martinez,et al.  Improved Heterogeneous Distance Functions , 1996, J. Artif. Intell. Res..

[17]  Yi Li,et al.  Boosting the K-Nearest-Neighborhood based incremental collaborative filtering , 2013, Knowl. Based Syst..

[18]  Yang Yu,et al.  Ensembling local learners ThroughMultimodal perturbation , 2005, IEEE Trans. Syst. Man Cybern. Part B.

[19]  Saso Dzeroski,et al.  Tree ensembles for predicting structured outputs , 2013, Pattern Recognit..

[20]  Bogdan Gabrys,et al.  Genetic algorithms in classifier fusion , 2006, Appl. Soft Comput..

[21]  Juan José Rodríguez Diez,et al.  An Experimental Study on Rotation Forest Ensembles , 2007, MCS.

[22]  Thiago J. M. Moura,et al.  Combining diversity measures for ensemble pruning , 2016, Pattern Recognit. Lett..

[23]  Loris Nanni,et al.  Particle swarm optimization for ensembling generation for evidential k-nearest-neighbour classifier , 2009, Neural Computing and Applications.

[24]  Krzysztof Jajuga,et al.  Fuzzy clustering with squared Minkowski distances , 2001, Fuzzy Sets Syst..

[25]  Juan José Rodríguez Diez,et al.  Rotation Forest: A New Classifier Ensemble Method , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[26]  Tin Kam Ho,et al.  The Random Subspace Method for Constructing Decision Forests , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[27]  Thierry Denoeux,et al.  An evidential classifier based on feature selection and two-step classification strategy , 2015, Pattern Recognit..

[28]  David L. Waltz,et al.  Toward memory-based reasoning , 1986, CACM.

[29]  Subhash C. Bagui,et al.  Combining Pattern Classifiers: Methods and Algorithms , 2005, Technometrics.

[30]  Qinghua Hu,et al.  Neighborhood classifiers , 2008, Expert Syst. Appl..

[31]  Qinghua Hu,et al.  Neighborhood rough set based heterogeneous feature subset selection , 2008, Inf. Sci..

[32]  Fajie Duan,et al.  Metric learning via feature weighting for scalable image retrieval , 2017, Pattern Recognit. Lett..

[33]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1997, EuroCOLT.

[34]  Janez Demsar,et al.  Statistical Comparisons of Classifiers over Multiple Data Sets , 2006, J. Mach. Learn. Res..

[35]  Fabio Roli,et al.  Designing multi-label classifiers that maximize F measures: State of the art , 2017, Pattern Recognit..

[36]  Bernard De Baets,et al.  Supervised distance metric learning through maximization of the Jeffrey divergence , 2017, Pattern Recognit..

[37]  Nicolás García-Pedrajas,et al.  Boosting k-nearest neighbor classifier by means of input space projection , 2009, Expert Syst. Appl..

[38]  Qinghua Hu,et al.  A weighted rough set based method developed for class imbalance learning , 2008, Inf. Sci..

[39]  Yiyu Yao,et al.  Relational Interpretations of Neigborhood Operators and Rough Set Approximation Operators , 1998, Inf. Sci..

[40]  E. M. Kleinberg,et al.  Stochastic discrimination , 1990, Annals of Mathematics and Artificial Intelligence.

[41]  Jack Sklansky,et al.  A note on genetic algorithms for large-scale feature selection , 1989, Pattern Recognit. Lett..

[42]  Tin Kam Ho,et al.  Nearest Neighbors in Random Subspaces , 1998, SSPR/SPR.

[43]  Naohiro Ishii,et al.  Combining Multiple k-Nearest Neighbor Classifiers Using Different Distance Functions , 2004, IDEAL.

[44]  Fang Liu,et al.  Random subspace based ensemble sparse representation , 2018, Pattern Recognit..