Predictive learning with structured (grouped) data

Many applications of machine learning involve sparse and heterogeneous data. For example, estimation of diagnostic models using patients' data from clinical studies requires effective integration of genetic, clinical and demographic data. Typically all heterogeneous inputs are properly encoded and mapped onto a single feature vector, used for estimating a classifier. This approach, known as standard inductive learning, is used in most application studies. Recently, several new learning methodologies have emerged. For instance, when training data can be naturally separated into several groups (or structured), we can view model estimation for each group as a separate task, leading to a Multi-Task Learning framework. Similarly, a setting where the training data are structured, but the objective is to estimate a single predictive model (for all groups), leads to the Learning with Structured Data and SVM+ methodology recently proposed by Vapnik [(2006). Empirical inference science afterword of 2006. Springer]. This paper describes a biomedical application of these new data modeling approaches for modeling heterogeneous data using several medical data sets. The characteristics of group variables are analyzed. Our comparisons demonstrate the advantages and limitations of these new approaches, relative to standard inductive SVM classifiers.

[1]  Rich Caruana,et al.  Multitask Learning , 1997, Machine Learning.

[2]  Masoud Nikravesh,et al.  Feature Extraction - Foundations and Applications , 2006, Feature Extraction.

[3]  Vladimir Cherkassky,et al.  Connection between SVM+ and multi-task learning , 2008, 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence).

[4]  Michael I. Jordan,et al.  Multi-task feature selection , 2006 .

[5]  Rajat Raina,et al.  Constructing informative priors using transfer learning , 2006, ICML.

[6]  Tom Heskes,et al.  Task Clustering and Gating for Bayesian Multitask Learning , 2003, J. Mach. Learn. Res..

[7]  Xuejun Liao,et al.  Radial Basis Function Network for Multitask Learning , 2005 .

[8]  Vladimir Vapnik,et al.  Learning using hidden information (Learning with teacher) , 2009, 2009 International Joint Conference on Neural Networks.

[9]  T. Ben-David,et al.  Exploiting Task Relatedness for Multiple , 2003 .

[10]  José Luis Rojo-Álvarez,et al.  Kernel Methods in Bioengineering, Signal And Image Processing , 2007 .

[11]  Vladimir Cherkassky,et al.  SVM+ regression and multi-task learning , 2009, 2009 International Joint Conference on Neural Networks.

[12]  L. Schoer,et al.  Learning for the Teacher. , 1970 .

[13]  Vladimir Vapnik,et al.  Statistical learning theory , 1998 .

[14]  Shai Ben-David,et al.  A theoretical framework for learning from a pool of disparate data sources , 2002, KDD.

[15]  Vladimir Cherkassky,et al.  Learning from data , 1998 .

[16]  Neil D. Lawrence,et al.  Learning to learn with the informative vector machine , 2004, ICML.

[17]  Massimiliano Pontil,et al.  Multi-Task Feature Learning , 2006, NIPS.

[18]  Massimiliano Pontil,et al.  Regularized multi--task learning , 2004, KDD.

[19]  Kristin P. Bennett,et al.  Constructing Orthogonal Latent Features for Arbitrary Loss , 2006, Feature Extraction.

[20]  Tong Zhang,et al.  A Framework for Learning Predictive Structures from Multiple Tasks and Unlabeled Data , 2005, J. Mach. Learn. Res..

[21]  M. Radmacher,et al.  Pitfalls in the use of DNA microarray data for diagnostic and prognostic classification. , 2003, Journal of the National Cancer Institute.

[22]  Vladimir Cherkassky,et al.  Application and development of new learning methodologies for fmri data analysis , 2007 .

[23]  Lawrence Carin,et al.  Radial Basis Function Network for Multi-task Learning , 2005, NIPS.

[24]  Vladimir Cherkassky,et al.  Learning Using Structured Data: Application to fMRI Data Analysis , 2007, 2007 International Joint Conference on Neural Networks.