Enhancing k-Nearest Neighbors through Learning Transformation Functions by Genetic Programming

The k-nearest neighbors algorithm (kNN) is renowned for solving classification tasks. The notion of kNN is to seek similar data instances in the dataset as prediction reference, for which the similarity between instances is ordinarily measured by Euclidean distance. Recently, some studies propose problem-tailored distance metrics to improve the classification performance of kNN. In this paper, we utilize genetic programming to learn the transformation function, which interprets the relationship of two data instances into a scalar differential. The differential of data pairs indicates the dissimilarity between two instances. This study considers two forms of transformation functions. Experimental results show the transform functions learned by GP can effectively enhance the performance of kNN.

[1]  Inderjit S. Dhillon,et al.  Metric and Kernel Learning Using a Linear Transformation , 2009, J. Mach. Learn. Res..

[2]  Anil K. Jain,et al.  Dimensionality reduction using genetic algorithms , 2000, IEEE Trans. Evol. Comput..

[3]  John R. Koza,et al.  Genetic programming 2 - automatic discovery of reusable programs , 1994, Complex Adaptive Systems.

[4]  Rong Jin,et al.  Distance Metric Learning: A Comprehensive Survey , 2006 .

[5]  Ahmed Bouridane,et al.  Simultaneous feature selection and feature weighting using Hybrid Tabu Search/K-nearest neighbor classifier , 2007, Pattern Recognit. Lett..

[6]  Michael I. Jordan,et al.  Distance Metric Learning with Application to Clustering with Side-Information , 2002, NIPS.

[7]  Rabab Kreidieh Ward,et al.  Genetic algorithms for feature selection and weighting, a review and study , 2001, Proceedings of Sixth International Conference on Document Analysis and Recognition.

[8]  Kilian Q. Weinberger,et al.  Distance Metric Learning for Large Margin Nearest Neighbor Classification , 2005, NIPS.

[9]  Tao Xiang,et al.  Learning to Compare: Relation Network for Few-Shot Learning , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[10]  Daniel Gutiérrez-Reina,et al.  Dissimilarity Metric Based on Local Neighboring Information and Genetic Programming for Data Dissemination in Vehicular Ad Hoc Networks (VANETs) , 2018, Sensors.

[11]  Alexandros Agapitos,et al.  Adaptive Distance Metrics for Nearest Neighbour Classification Based on Genetic Programming , 2013, EuroGP.

[12]  Mengjie Zhang,et al.  Genetic programming for feature construction and selection in classification on high-dimensional data , 2016, Memetic Comput..

[13]  Yann LeCun,et al.  Efficient Pattern Recognition Using a New Transformation Distance , 1992, NIPS.

[14]  Asoke K. Nandi,et al.  Extending genetic programming for multi-class classification by combining K-nearest neighbor , 2005, Proceedings. (ICASSP '05). IEEE International Conference on Acoustics, Speech, and Signal Processing, 2005..

[15]  Peter E. Hart,et al.  Nearest neighbor pattern classification , 1967, IEEE Trans. Inf. Theory.

[16]  Dimitrios Gunopulos,et al.  Locally Adaptive Metric Nearest-Neighbor Classification , 2002, IEEE Trans. Pattern Anal. Mach. Intell..

[17]  Brian Kulis,et al.  Metric Learning: A Survey , 2013, Found. Trends Mach. Learn..

[18]  Asoke K. Nandi,et al.  Automatic digital modulation classification using Genetic Programming with K-Nearest Neighbor , 2010, 2010 - MILCOM 2010 MILITARY COMMUNICATIONS CONFERENCE.

[19]  A. Asuncion,et al.  UCI Machine Learning Repository, University of California, Irvine, School of Information and Computer Sciences , 2007 .