A Novel Multiple Kernel Learning Method Based on the Kullback–Leibler Divergence

In this paper we develop a novel multiple kernel learning (MKL) model that is based on the idea of the multiplicative perturbation of data in a new feature space in the framework of uncertain convex programs (UCPs). In the proposed model, we utilize the Kullback–Leibler divergence to measure the difference between the estimated kernel weights and ideal kernel weights. Instead of directly handling the proposed model in the primal, we obtain the optimistic counterpart of its Langrage dual in terms of the theory of UCPs and solve it by using the alternating optimization. In the case of a varying parameter, the proposed model gives the solution path from a robust combined kernel to some combined kernel corresponding to the initially ideal kernel weights. In addition, we also give a simple strategy to select the initial kernel weights as the ideal kernel weights if any prior knowledge of kernel weights is not available. Experimental results on several data sets show that the proposed model can obtain competitive performance with some of the state-of-the-art MKL algorithms.

[1]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[2]  Zenglin Xu,et al.  Efficient Sparse Generalized Multiple Kernel Learning , 2011, IEEE Transactions on Neural Networks.

[3]  R. Ravi,et al.  An FPTAS for minimizing a class of low-rank quasi-concave functions over a convex set , 2013, Oper. Res. Lett..

[4]  K. Schittkowski,et al.  NONLINEAR PROGRAMMING , 2022 .

[5]  Nello Cristianini,et al.  Learning the Kernel Matrix with Semidefinite Programming , 2002, J. Mach. Learn. Res..

[6]  Yoram Singer,et al.  Efficient projections onto the l1-ball for learning in high dimensions , 2008, ICML '08.

[7]  Francis R. Bach,et al.  Consistency of the group Lasso and multiple kernel learning , 2007, J. Mach. Learn. Res..

[8]  Donald Goldfarb,et al.  Second-order cone programming , 2003, Math. Program..

[9]  Ivor W. Tsang,et al.  This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 1 Soft Margin Multiple Kernel Learning , 2022 .

[10]  Amir Beck,et al.  Duality in robust optimization: Primal worst equals dual best , 2009, Oper. Res. Lett..

[11]  Michael I. Jordan,et al.  Multiple kernel learning, conic duality, and the SMO algorithm , 2004, ICML.

[12]  Vaithilingam Jeyakumar,et al.  Strong Duality in Robust Convex Programming: Complete Characterizations , 2010, SIAM J. Optim..

[13]  Zenglin Xu,et al.  Simple and Efficient Multiple Kernel Learning by Group Lasso , 2010, ICML.

[14]  Chiranjib Bhattacharyya,et al.  Efficient methods for robust classification under uncertainty in kernel matrices , 2012, J. Mach. Learn. Res..

[15]  Gunnar Rätsch,et al.  Large Scale Multiple Kernel Learning , 2006, J. Mach. Learn. Res..

[16]  Shixiong Xia,et al.  Training Lp norm multiple kernel learning in the primal , 2013, Neural Networks.

[17]  Guoyin Li,et al.  Support vector machine classifiers with uncertain knowledge sets via robust optimization , 2014 .

[18]  Zenglin Xu,et al.  Smooth Optimization for Effective Multiple Kernel Learning , 2010, AAAI.

[19]  Dimitri P. Bertsekas,et al.  Nonlinear Programming , 1997 .

[20]  Janez Demsar,et al.  Statistical Comparisons of Classifiers over Multiple Data Sets , 2006, J. Mach. Learn. Res..

[21]  M. Kloft,et al.  l p -Norm Multiple Kernel Learning , 2011 .

[22]  S. V. N. Vishwanathan,et al.  Multiple Kernel Learning and the SMO Algorithm , 2010, NIPS.

[23]  Ethem Alpaydin,et al.  Multiple Kernel Learning Algorithms , 2011, J. Mach. Learn. Res..

[24]  Yves Grandvalet,et al.  Composite kernel learning , 2008, ICML '08.

[25]  Laurent El Ghaoui,et al.  Robust Optimization , 2021, ICORES.

[26]  Vaithilingam Jeyakumar,et al.  Robust conjugate duality for convex optimization under uncertainty with application to data classifi , 2011 .

[27]  Chiranjib Bhattacharyya,et al.  Variable Sparsity Kernel Learning , 2011, J. Mach. Learn. Res..