Scaling KNN multi-class twin support vector machine via safe instance reduction

Abstract k-nearest neighbor-based weighted multi-class twin support vector machine(KMTSVM) is an effective algorithm to deal with multi-class optimization problems. The superiority of KMTSVM is that it adopts a “1-versus-1-versus-rest” structure, and takes the distribution information of all the instances into consideration. However, it costs much time to handle large scale problems. Motivated by the sparse solution of KMTSVM, in this paper, we propose a safe instance reduction rule to improve its computational efficiency, termed as SIR-KMTSVM. The SIR-KMTSVM can delete a majority of redundant instances both for focused classes and remaining classes, then the operation speed can be accelerated greatly. And our instance reduction rule is safe in the sense that the reduced problem can derive an identical optimal solution as the original one. More importantly, we analysis that the different k-nearest neighbors will have different acceleration effect on our SIR-KMTSVM. Besides, a fast algorithm DCDM is introduced to handle relatively large scale problems more efficiently. Sequential versions of SIR-KMTSVM are further introduced to substantially accelerate the whole training process. Experimental results on an artificial dataset, seventeen benchmark datasets and a real dataset confirm the effectiveness of our proposed algorithm.

[1]  Rui Guo,et al.  A Twin Multi-Class Classification Support Vector Machine , 2012, Cognitive Computation.

[2]  Andreu Català,et al.  K-SVCR. A support vector machine for multi-class classification , 2003, Neurocomputing.

[3]  Chih-Jen Lin,et al.  A comparison of methods for multiclass support vector machines , 2002, IEEE Trans. Neural Networks.

[4]  Yitian Xu,et al.  K-nearest neighbor-based weighted multi-class twin support vector machine , 2016, Neurocomputing.

[5]  Bernhard Schölkopf,et al.  Sampling Techniques for Kernel Methods , 2001, NIPS.

[6]  Shuicheng Yan,et al.  Graph embedding: a general framework for dimensionality reduction , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[7]  Xiao-Jun Wu,et al.  Multiple birth least squares support vector machine for multi-class classification , 2017, Int. J. Mach. Learn. Cybern..

[8]  Jiang Zhao,et al.  A safe sample screening rule for Universum support vector machines , 2017, Knowl. Based Syst..

[9]  Peter E. Hart,et al.  Nearest neighbor pattern classification , 1967, IEEE Trans. Inf. Theory.

[10]  Jieping Ye,et al.  Scaling SVM and Least Absolute Deviations via Exact Data Reduction , 2013, ICML.

[11]  Z. Ying,et al.  A simple resampling method by perturbing the minimand , 2001 .

[12]  Chih-Jen Lin,et al.  A dual coordinate descent method for large-scale linear SVM , 2008, ICML '08.

[13]  Shangbing Gao,et al.  Weighted Twin Support Vector Machines with Local Information and its application , 2012, Neural Networks.

[14]  Bin Chen,et al.  Proximal support vector machine using local information , 2009, Neurocomputing.

[15]  Jason Weston,et al.  Support vector machines for multi-class pattern recognition , 1999, ESANN.

[16]  Thorsten Joachims,et al.  Text Categorization with Support Vector Machines: Learning with Many Relevant Features , 1998, ECML.

[17]  Jiawei Han,et al.  Classifying large data sets using SVMs with hierarchical clusters , 2003, KDD '03.

[18]  Xianli Pan,et al.  Safe Screening Rules for Accelerating Twin Support Vector Machine Classification , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[19]  Daniel Boley,et al.  On Approximate Solutions to Support Vector Machines , 2006, SDM.

[20]  Chih-Jen Lin,et al.  Trust region Newton methods for large-scale logistic regression , 2007, ICML '07.

[21]  Ronald R. Coifman,et al.  Data Fusion and Multicue Data Matching by Diffusion Maps , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[22]  Yuh-Jye Lee,et al.  RSVM: Reduced Support Vector Machines , 2001, SDM.

[23]  David Preiss,et al.  Gateaux differentiable functions are somewhere Frechet differentiable , 1984 .

[24]  Moshe Buchinsky Recent Advances in Quantile Regression Models: A Practical Guideline for Empirical Research , 1998 .

[25]  Yuan-Hai Shao,et al.  Multiple recursive projection twin support vector machine for multi-class classification , 2014, International Journal of Machine Learning and Cybernetics.

[26]  Xianli Pan,et al.  Two effective sample selection methods for support vector machine , 2016, J. Intell. Fuzzy Syst..

[27]  Yitian Xu,et al.  K-nearest neighbor-based weighted twin support vector regression , 2014, Applied Intelligence.

[28]  John C. Platt,et al.  Fast training of support vector machines using sequential minimal optimization, advances in kernel methods , 1999 .

[29]  Koby Crammer,et al.  On the Algorithmic Implementation of Multiclass Kernel-based Vector Machines , 2002, J. Mach. Learn. Res..

[30]  Leonardo Franco,et al.  Multiclass Pattern Recognition Extension for the New C-Mantec Constructive Neural Network Algorithm , 2010, Cognitive Computation.

[31]  Osman Güler,et al.  Foundations of Optimization , 2010 .

[32]  Zhiquan Qi,et al.  A New Support Vector Machine for Multi-class Classification , 2005, CIT.

[33]  Yitian Xu,et al.  A safe screening rule for Laplacian support vector machine , 2018, Eng. Appl. Artif. Intell..

[34]  Robert Tibshirani,et al.  Classification by Pairwise Coupling , 1997, NIPS.

[35]  Xianli Pan,et al.  K-nearest neighbor based structural twin support vector machine , 2015, Knowl. Based Syst..

[36]  Federico Girosi,et al.  Training support vector machines: an application to face detection , 1997, Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[37]  Jia Yu,et al.  KNN-based weighted rough ν-twin support vector machine , 2014, Knowl. Based Syst..

[38]  Zhirong Sun,et al.  Support vector machine approach for protein subcellular localization prediction , 2001, Bioinform..

[39]  Rui Guo,et al.  An improved ν-twin support vector machine , 2013, Applied Intelligence.

[40]  Laurent El Ghaoui,et al.  Safe Feature Elimination in Sparse Supervised Learning , 2010, ArXiv.

[41]  Thorsten Joachims,et al.  Training linear SVMs in linear time , 2006, KDD '06.

[42]  Ichiro Takeuchi,et al.  Safe Screening of Non-Support Vectors in Pathwise SVM Computation , 2013, ICML.