Multi-class kernel margin maximization for kernel learning

Two-stage multiple kernel learning (MKL) algorithms have been intensively studied due to its high efficiency and effectiveness. Pioneering work on this regard attempts to optimize the combination coefficients by maximizing the multi-class margin of a kernel, while obtaining unsatisfying performance. In this paper, we attribute this poor performance to the way in calculating the multi-class margin of a kernel. In specific, we argue that for each sample only the k-nearest neighbors, while not all samples with the same label, should be selected for calculating the margin. After that, we also develop another sparse variant which is able to automatically identify the nearest neighbors and the corresponding weights of each sample. Extensive experimental results on ten UCI data sets and six MKL benchmark data sets demonstrate the effectiveness and efficiency of the proposed algorithms.

[1]  Lei Wang,et al.  An Efficient Approach to Integrating Radius Information into Multiple Kernel Learning , 2013, IEEE Transactions on Cybernetics.

[2]  Theodoros Damoulas,et al.  Probabilistic multi-class multi-kernel learning: on protein fold recognition and remote homology detection , 2008, Bioinform..

[3]  Josef Kittler,et al.  Non-Sparse Multiple Kernel Fisher Discriminant Analysis , 2012, J. Mach. Learn. Res..

[4]  Zhiyuan Li,et al.  Quasi-newton method for LP multiple kernel learning , 2016, Neurocomputing.

[5]  Jieping Ye,et al.  Multi-class Discriminant Kernel Learning via Convex Programming , 2008, J. Mach. Learn. Res..

[6]  Nello Cristianini,et al.  Learning the Kernel Matrix with Semidefinite Programming , 2002, J. Mach. Learn. Res..

[7]  Lei Wang,et al.  Multiple kernel extreme learning machine , 2015, Neurocomputing.

[8]  Mehryar Mohri,et al.  Multi-Class Classification with Maximum Margin Multiple Kernel , 2013, ICML.

[9]  Lei Wang,et al.  Sample-Adaptive Multiple Kernel Learning , 2014, AAAI.

[10]  Jianping Yin,et al.  Incorporation of radius-info can be simple with SimpleMKL , 2012, Neurocomputing.

[11]  Rong Jin,et al.  Multiple Kernel Learning for Visual Object Recognition: A Review , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[12]  Zenglin Xu,et al.  Simple and Efficient Multiple Kernel Learning by Group Lasso , 2010, ICML.

[13]  Yoram Singer,et al.  Pegasos: primal estimated sub-gradient solver for SVM , 2007, ICML '07.

[14]  Marius Kloft,et al.  Learning Kernels Using Local Rademacher Complexity , 2013, NIPS.

[15]  Feiping Nie,et al.  A general kernelization framework for learning algorithms based on kernel PCA , 2010, Neurocomputing.

[16]  K JainAnil,et al.  Multiple Kernel Learning for Visual Object Recognition , 2014 .

[17]  Wenhua Wang,et al.  Classification by semi-supervised discriminative regularization , 2010, Neurocomputing.

[18]  Cheng Soon Ong,et al.  Multiclass multiple kernel learning , 2007, ICML '07.

[19]  Ethem Alpaydin,et al.  Multiple Kernel Learning Algorithms , 2011, J. Mach. Learn. Res..

[20]  Koray Kavukcuoglu,et al.  A Binary Classification Framework for Two-Stage Multiple Kernel Learning , 2012, ICML.

[21]  Andrew Zisserman,et al.  A Visual Vocabulary for Flower Classification , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).

[22]  Mehryar Mohri,et al.  Algorithms for Learning Kernels Based on Centered Alignment , 2012, J. Mach. Learn. Res..

[23]  Mehmet G nen Bayesian Efficient Multiple Kernel Learning , 2012, ICML 2012.