FKNDT: A Flexible Kernel by Negotiating Between Data-dependent Kernel Learning and Task-dependent Kernel Learning

Kernel learning is a challenging issue which has been vastly investigated over the last decades. The performance of kernel-based methods broadly relies on selecting an appropriate kernel. In machine learning community, a fundamental problem is how to model a suitable kernel. The traditional kernels, e.g., Gaussian kernel and polynomial kernel, are not adequately flexible to employ the information of the given data. Classical kernels are unable to sufficiently depict the characteristics of data similarities. To alleviate this problem, this paper presents a Flexible Kernel by Negotiating between Data-dependent kernel learning and Task-dependent kernel learning termed as FKNDT. Our method learns a suitable kernel by way of the Hadamard product of two types of kernels; a data-dependent kernel and a set of pre-specified classical kernels as a task-dependent kernel. We evaluate a flexible kernel in a supervised manner via Support Vector Machines (SVM). We model a learning process as a joint optimization problem including data-dependent kernel matrix learning, multiple kernel learning by means of quadratic programming, and standard SVM optimization. The experimental results demonstrate our technique provides a more effective kernel than the traditional kernels. Our method is better than other state-of-the-art kernel-based algorithms in terms of classification accuracy on fifteen benchmark datasets.

[1]  Ethem Alpaydin,et al.  Multiple Kernel Learning Algorithms , 2011, J. Mach. Learn. Res..

[2]  Peyman Milanfar,et al.  Kernel Regression for Image Processing and Reconstruction , 2007, IEEE Transactions on Image Processing.

[3]  Tomaso A. Poggio,et al.  Regularization Networks and Support Vector Machines , 2000, Adv. Comput. Math..

[4]  Bernhard Schölkopf,et al.  Measuring Statistical Dependence with Hilbert-Schmidt Norms , 2005, ALT.

[5]  Andrew Gordon Wilson,et al.  Deep Kernel Learning , 2015, AISTATS.

[6]  Inderjit S. Dhillon,et al.  Low-Rank Kernel Learning with Bregman Matrix Divergences , 2009, J. Mach. Learn. Res..

[7]  Inderjit S. Dhillon,et al.  Geometry-aware metric learning , 2009, ICML '09.

[8]  Charles A. Micchelli,et al.  Learning Convex Combinations of Continuously Parameterized Basic Kernels , 2005, COLT.

[9]  Yong Liu,et al.  Infinite Kernel Learning: Generalization Bounds and Algorithms , 2017, AAAI.

[10]  Anthony Widjaja,et al.  Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond , 2003, IEEE Transactions on Neural Networks.

[11]  Lei Wang,et al.  An Adaptive Approach to Learning Optimal Neighborhood Kernels , 2013, IEEE Transactions on Cybernetics.

[12]  Ivor W. Tsang,et al.  Parameter-Free Spectral Kernel Learning , 2010, UAI.

[13]  Rama Chellappa,et al.  Cross-Sensor Iris Recognition through Kernel Learning , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[14]  Mehmet G nen Bayesian Efficient Multiple Kernel Learning , 2012, ICML 2012.

[15]  Rong Jin,et al.  Multiple Kernel Learning for Visual Object Recognition: A Review , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[16]  J. Malick,et al.  Projection methods in conic optimization , 2011, 1103.1511.

[17]  John C. Duchi,et al.  Learning Kernels with Random Features , 2016, NIPS.

[18]  Michael H. Bowling,et al.  Alignment based kernel learning with a continuous set of base kernels , 2011, Machine Learning.

[19]  Larry S. Davis,et al.  Incremental Multiple Kernel Learning for object recognition , 2009, 2009 IEEE 12th International Conference on Computer Vision.

[20]  Stephen Lin,et al.  Graph Embedding and Extensions: A General Framework for Dimensionality Reduction , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[21]  Lei Wang,et al.  Multiple Kernel k-Means Clustering with Matrix-Induced Regularization , 2016, AAAI.

[22]  Ivor W. Tsang,et al.  Two-Layer Multiple Kernel Learning , 2011, AISTATS.

[23]  Viet Hoai Vo,et al.  Multiple Modal Features and Multiple Kernel Learning for Human Daily Activity Recognition , 2018, Science and Technology Development Journal.

[24]  Inderjit S. Dhillon,et al.  Metric and Kernel Learning Using a Linear Transformation , 2009, J. Mach. Learn. Res..

[25]  Lawrence K. Saul,et al.  Kernel Methods for Deep Learning , 2009, NIPS.

[26]  Alexander J. Smola,et al.  Support Vector Regression Machines , 1996, NIPS.

[27]  Mehmet Gönen,et al.  Bayesian Efficient Multiple Kernel Learning , 2012, ICML.

[28]  Jie Yang,et al.  Nonlinear Pairwise Layer and Its Training for Kernel Learning , 2018, AAAI.

[29]  Jinfeng Yi,et al.  Stochastic Optimization for Kernel PCA , 2016, AAAI.

[30]  Cristian Sminchisescu,et al.  Kernel Learning by Unconstrained Optimization , 2009, AISTATS.

[31]  Xiaobo Zhou,et al.  Incremental Kernel Ridge Regression for the Prediction of Soft Tissue Deformations , 2012, MICCAI.

[32]  Inderjit S. Dhillon,et al.  Kernel k-means: spectral clustering and normalized cuts , 2004, KDD.

[33]  Erik Cambria,et al.  Deep Convolutional Neural Network Textual Features and Multiple Kernel Learning for Utterance-level Multimodal Sentiment Analysis , 2015, EMNLP.

[34]  Karthikeyan Natesan Ramamurthy,et al.  Multiple Kernel Sparse Representations for Supervised and Unsupervised Learning , 2013, IEEE Transactions on Image Processing.

[35]  Nello Cristianini,et al.  Learning the Kernel Matrix with Semidefinite Programming , 2002, J. Mach. Learn. Res..

[36]  Mehryar Mohri,et al.  Algorithms for Learning Kernels Based on Centered Alignment , 2012, J. Mach. Learn. Res..

[37]  Binbin Pan,et al.  A Novel Framework for Learning Geometry-Aware Kernels , 2016, IEEE Transactions on Neural Networks and Learning Systems.

[38]  J. Platt Sequential Minimal Optimization : A Fast Algorithm for Training Support Vector Machines , 1998 .

[39]  Jing Li,et al.  Heterogeneous data fusion for alzheimer's disease study , 2008, KDD.

[40]  Mehran Kafai,et al.  CROification: Accurate Kernel Classification with the Efficiency of Sparse Linear SVM , 2019, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[41]  Hongdong Li,et al.  Expanding the Family of Grassmannian Kernels: An Embedding Perspective , 2014, ECCV.