Multiple kernel learning based on local and nonlinear combinations

In recent years, different methods based on kernels have been used with success in a variety of tasks such as classification. However, in the typical use of these methods, the choice of the optimal kernel is crucial to improve the performance of a specific task. So, instead of selecting a single kernel, multiple kernel learning (MKL) has been proposed which uses a combination of kernels, where the weight of each kernel is optimized in the training stage. MKL methods use kernels in linear, nonlinear or data-dependent combinations. Methods based on MKL have performed better than methods using a single kernel such as the Support Vector Machine (SVM). In this article, we propose a new MKL method, which is based on a local (data dependent) and nonlinear combination of different kernels using a gating model for selecting the appropriate kernel function. We call our proposal as localized nonlinear multiple kernel learning (LNLMKL). In our experiments for binary microarray classification, different kernels were used in SVM and different kernels combinations were used for our proposal and for the other MKL methods. Finally, we report the results of these experiments using eight high-dimensional microarray data sets demonstrating that our proposal have performed better than the other methods analyzed.

[1]  Jean YH Yang,et al.  Bioconductor: open software development for computational biology and bioinformatics , 2004, Genome Biology.

[2]  U. Alon,et al.  Broad patterns of gene expression revealed by clustering analysis of tumor and normal colon tissues probed by oligonucleotide arrays. , 1999, Proceedings of the National Academy of Sciences of the United States of America.

[3]  E. Petricoin,et al.  Use of proteomic patterns in serum to identify ovarian Cancer , 2002 .

[4]  Mehryar Mohri,et al.  Learning Non-Linear Combinations of Kernels , 2009, NIPS.

[5]  Sayan Mukherjee,et al.  Classifying Microarray Data Using Support Vector Machines , 2003 .

[6]  Natalia Shulzhenko,et al.  Microarrays for cancer diagnosis and classification. , 2007, Advances in experimental medicine and biology.

[7]  S. Knudsen,et al.  Guide to Analysis of DNA Microarray Data: Knudsen/DNA Microarray 2e , 2005 .

[8]  Alneu de Andrade Lopes,et al.  Link Prediction in Complex Networks Based on Cluster Information , 2012, SBIA.

[9]  Si Wu,et al.  Improving support vector machine classifiers by modifying kernel functions , 1999, Neural Networks.

[10]  P. Hall,et al.  An expression signature for p53 status in human breast cancer predicts mutation status, transcriptional effects, and patient survival. , 2005, Proceedings of the National Academy of Sciences of the United States of America.

[11]  Bernhard E. Boser,et al.  A training algorithm for optimal margin classifiers , 1992, COLT '92.

[12]  Ethem Alpaydin,et al.  Localized multiple kernel learning , 2008, ICML '08.

[13]  Yves Grandvalet,et al.  More efficiency in multiple kernel learning , 2007, ICML '07.

[14]  Umberto Castellani,et al.  Combining Data Sources Nonlinearly for Cell Nucleus Classification of Renal Cell Carcinoma , 2011, SIMBAD.

[15]  Ethem Alpaydin,et al.  Localized algorithms for multiple kernel learning , 2013, Pattern Recognit..

[16]  Michael I. Jordan,et al.  Multiple kernel learning, conic duality, and the SMO algorithm , 2004, ICML.

[17]  Vladimir Vapnik,et al.  Statistical learning theory , 1998 .

[18]  David Haussler,et al.  Proceedings of the fifth annual workshop on Computational learning theory , 1992, COLT 1992.

[19]  Gunnar Rätsch,et al.  Large Scale Multiple Kernel Learning , 2006, J. Mach. Learn. Res..

[20]  Robert A. Jacobs,et al.  Hierarchical Mixtures of Experts and the EM Algorithm , 1993, Neural Computation.

[21]  Nello Cristianini,et al.  A statistical framework for genomic data fusion , 2004, Bioinform..

[22]  J. Mehta Gene expression analysis in breast cancer , 2010 .

[23]  J. Mesirov,et al.  Molecular classification of cancer: class discovery and class prediction by gene expression monitoring. , 1999, Science.

[24]  L. Holmberg,et al.  Gene expression profiling spares early breast cancer patients from adjuvant therapy: derived and validated in two population-based cohorts , 2005, Breast Cancer Research.

[25]  Harold W. Kuhn,et al.  Nonlinear programming: a historical view , 1982, SMAP.

[26]  Nello Cristianini,et al.  Learning the Kernel Matrix with Semidefinite Programming , 2002, J. Mach. Learn. Res..

[27]  Paulo J. G. Lisboa,et al.  Application notes: data mining in cancer research , 2010 .

[28]  P. Hall,et al.  An expression signature for p 53 status in human breast cancer predicts mutation status , transcriptional effects , and patient survival , 2005 .

[29]  Yudong D. He,et al.  Gene expression profiling predicts clinical outcome of breast cancer , 2002, Nature.

[30]  J. Bergh,et al.  Strong Time Dependence of the 76-Gene Prognostic Signature for Node-Negative Breast Cancer Patients in the TRANSBIG Multicenter Independent Validation Series , 2007, Clinical Cancer Research.

[31]  Ethem Alpaydin,et al.  Multiple Kernel Learning Algorithms , 2011, J. Mach. Learn. Res..

[32]  P. F. Macgregor,et al.  Application of microarrays to the analysis of gene expression in cancer. , 2002, Clinical chemistry.

[33]  Janez Demsar,et al.  Statistical Comparisons of Classifiers over Multiple Data Sets , 2006, J. Mach. Learn. Res..

[34]  Shweta Seeja K.R. Microarray Data Classification Using Support Vector Machine , 2011 .

[35]  Steen Knudsen,et al.  Guide to analysis of DNA microarray data , 2004 .

[36]  K. Cobb In a Vast Sea of Data , 2022 .

[37]  Chih-Jen Lin,et al.  LIBSVM: A library for support vector machines , 2011, TIST.

[38]  J. Foekens,et al.  Gene-expression profiles to predict distant metastasis of lymph-node-negative primary breast cancer , 2005, The Lancet.