A generalized weighted distance k-Nearest Neighbor for multi-label problems

Abstract In multi-label classification, each instance is associated with a set of pre-specified labels. One common approach is to use Binary Relevance (BR) paradigm to learn each label by a base classifier separately. Use of k-Nearest Neighbor (kNN) as the base classifier (denoted as BRkNN) is a simple, descriptive and powerful approach. In binary relevance a highly imbalanced view of dataset is used. However, kNN is known to perform poorly on imbalanced data. One approach to deal with this is to define the distance function in a parametric form and use the training data to adjust the parameters (i.e. adjusting boundaries between classes) by optimizing a performance measure customized for imbalanced data e.g. F-measure. Prototype Weighting (PW) scheme presented in the literature (Paredes & Vidal, 2006) uses gradient descent to specify the parameters by minimizing the classification error-rate on training data. This paper presents a generalized version of PW. First, instead of minimizing the error-rate proposed in PW, the generalized PW supports also other objective functions that use elements of confusion matrix (including F-measure). Second, PW originally presented for 1NN is extended to the general case of kNN (i.e., k > = 1 ). For problems having highly overlapped classes, it is expected to perform better since a value of k > 1 produces smoother decision boundaries which in turn can improve generalization. In multi-label problems with many labels or problems with highly overlapped classes, the proposed generalized PW is expected to significantly improve the performance as it involves many decision boundaries. The performance of the proposed method has been compared with state-of-the-art methods in multi-label classification containing 6 lazy classifiers based on kNN. Experiments show that the proposed method significantly outperforms other methods.

[1]  Enrique Vidal,et al.  Learning weighted metrics to minimize nearest-neighbor classification error , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[2]  Grigorios Tsoumakas,et al.  Dealing with Concept Drift and Class Imbalance in Multi-Label Stream Classification , 2011, IJCAI.

[3]  Fabio Roli,et al.  Threshold optimisation for multi-label classifiers , 2013, Pattern Recognit..

[4]  Yongzhao Zhan,et al.  Improved pseudo nearest neighbor classification , 2014, Knowl. Based Syst..

[5]  Thierry Denoeux,et al.  Multi-label classification algorithm derived from K-nearest neighbor rule with label dependencies , 2008, 2008 16th European Signal Processing Conference.

[6]  Enrique Vidal,et al.  Learning prototypes and distances (LPD). A prototype reduction technique based on nearest neighbor error minimization , 2004, Proceedings of the 17th International Conference on Pattern Recognition, 2004. ICPR 2004..

[7]  Peter E. Hart,et al.  Nearest neighbor pattern classification , 1967, IEEE Trans. Inf. Theory.

[8]  Min-Ling Zhang,et al.  A Review on Multi-Label Learning Algorithms , 2014, IEEE Transactions on Knowledge and Data Engineering.

[9]  Grigorios Tsoumakas,et al.  An Empirical Study of Lazy Multilabel Classification Algorithms , 2008, SETN.

[10]  Hong Liu,et al.  Coarse to fine K nearest neighbor classifier , 2013, Pattern Recognit. Lett..

[11]  Zhi-Hua Zhou,et al.  ML-KNN: A lazy learning approach to multi-label learning , 2007, Pattern Recognit..

[12]  Sebastián Ventura,et al.  Multi‐label learning: a review of the state of the art and ongoing research , 2014, WIREs Data Mining Knowl. Discov..

[13]  Mansoor Zolghadri Jahromi,et al.  Multi-label classification systems by the use of supervised clustering , 2017, 2017 Artificial Intelligence and Signal Processing Conference (AISP).

[14]  Geoff Holmes,et al.  Classifier chains for multi-label classification , 2009, Machine Learning.

[15]  Jianhua Xu,et al.  Multi-Label Weighted k-Nearest Neighbor Classifier with Adaptive Weight Estimation , 2011, ICONIP.

[16]  Tibério S. Caetano,et al.  Reverse Multi-Label Learning , 2010, NIPS.

[17]  Josef Kittler,et al.  Inverse random under sampling for class imbalance problem and its application to multi-label classification , 2012, Pattern Recognit..

[18]  Francisco Charte,et al.  Addressing imbalance in multilabel classification: Measures and random resampling algorithms , 2015, Neurocomputing.

[19]  José Ramón Quevedo,et al.  Multilabel classifiers with a probabilistic thresholding strategy , 2012, Pattern Recognit..

[20]  José Cristóbal Riquelme Santos,et al.  Improving the k-Nearest Neighbour Rule by an Evolutionary Voting Approach , 2014, HAIS.

[21]  José Cristóbal Riquelme Santos,et al.  On the evolutionary weighting of neighbours and features in the k-nearest neighbour rule , 2017, Neurocomputing.

[22]  Yingjie Tian,et al.  Joint Ranking SVM and Binary Relevance with Robust Low-Rank Learning for Multi-Label Classification , 2019, Neural Networks.

[23]  Lu Sun,et al.  Simultaneous Nonlinear Label-Instance Embedding for Multi-label Classification , 2016, S+SSPR.

[24]  Grigorios Tsoumakas,et al.  Random k -Labelsets: An Ensemble Method for Multilabel Classification , 2007, ECML.

[25]  Jianping Gou,et al.  A Novel Weighted Voting for K-Nearest Neighbor Rule , 2011, J. Comput..

[26]  Eyke Hüllermeier,et al.  Optimizing the F-Measure in Multi-Label Classification: Plug-in Rule Approach versus Structured Loss Minimization , 2013, ICML.

[27]  Sebastián Ventura,et al.  Evolutionary feature weighting to improve the performance of multi-label lazy algorithms , 2014, Integr. Comput. Aided Eng..

[28]  Yongzhao Zhan,et al.  A Local Mean Representation-based K-Nearest Neighbor Classifier , 2019, ACM Trans. Intell. Syst. Technol..

[29]  Baoyuan Wu,et al.  Constrained Submodular Minimization for Missing Labels and Class Imbalance in Multi-label Learning , 2016, AAAI.

[30]  Lu Sun,et al.  Multi-label classification with meta-label-specific features , 2016, 2016 23rd International Conference on Pattern Recognition (ICPR).

[31]  Farzana Kabir Ahmad,et al.  Trending Challenges in Multi Label Classification , 2016 .

[32]  Leon N. Cooper,et al.  Improving nearest neighbor rule with a simple adaptive distance measure , 2007, Pattern Recognit. Lett..

[33]  Grigorios Tsoumakas,et al.  Multi-Label Classification: An Overview , 2007, Int. J. Data Warehous. Min..

[34]  Sahibsingh A. Dudani The Distance-Weighted k-Nearest-Neighbor Rule , 1976, IEEE Transactions on Systems, Man, and Cybernetics.

[35]  Grigorios Tsoumakas,et al.  Mining Multi-label Data , 2010, Data Mining and Knowledge Discovery Handbook.

[36]  Sunita Sarawagi,et al.  Discriminative Methods for Multi-labeled Classification , 2004, PAKDD.

[37]  Robert Ivor John,et al.  A method of learning weighted similarity function to improve the performance of nearest neighbor , 2009, Inf. Sci..

[38]  Trevor Hastie,et al.  The Elements of Statistical Learning , 2001 .

[39]  Eyke Hüllermeier,et al.  Combining instance-based learning and logistic regression for multilabel classification , 2009, Machine Learning.

[40]  Jianping Gou,et al.  A new distance-weighted k-nearest neighbor classifier , 2012 .

[41]  Sarah Vluymans,et al.  Multi-label classification using a fuzzy rough neighborhood consensus , 2018, Inf. Sci..

[42]  Yong Zeng,et al.  Pseudo nearest neighbor rule for pattern classification , 2009, Expert Syst. Appl..

[43]  Jesús Alcalá-Fdez,et al.  KEEL Data-Mining Software Tool: Data Set Repository, Integration of Algorithms and Experimental Analysis Framework , 2011, J. Multiple Valued Log. Soft Comput..

[44]  Grigorios Tsoumakas,et al.  MULAN: A Java Library for Multi-Label Learning , 2011, J. Mach. Learn. Res..

[45]  Mohammad S. Sorower A Literature Survey on Algorithms for Multi-label Learning , 2010 .

[46]  Francisco Charte,et al.  MLSMOTE: Approaching imbalanced multilabel learning through synthetic instance generation , 2015, Knowl. Based Syst..

[47]  Iain D. Currie,et al.  Distribution‐Free Tests , 1989 .

[48]  Yoshihiko Hamamoto,et al.  A local mean-based nonparametric classifier , 2006, Pattern Recognit. Lett..