Supervised patient similarity measure of heterogeneous patient records

Patient similarity assessment is an important task in the context of patient cohort identif cation for comparative effectiveness studies and clinical decision support applications. The goal is to derive clinically meaningful distance metric to measure the similarity between patients represented by their key clinical indicators. How to incorporate physician feedback with regard to the retrieval results? How to interactively update the underlying similarity measure based on the feedback? Moreover, often different physicians have different understandings of patient similarity based on their patient cohorts. The distance metric learned for each individual physician often leads to a limited view of the true underlying distance metric. How to integrate the individual distance metrics from each physician into a globally consistent unif ed metric? We describe a suite of supervised metric learning approaches that answer the above questions. In particular, we present Locally Supervised Metric Learning (LSML) to learn a generalized Mahalanobis distance that is tailored toward physician feedback. Then we describe the interactive metric learning (iMet) method that can incrementally update an existing metric based on physician feedback in an online fashion. To combine multiple similarity measures from multiple physicians, we present Composite Distance Integration (Comdi) method. In this approach we f rst construct discriminative neighborhoods from each individual metrics, then combine them into a single optimal distance metric. Finally, we present a clinical decision support prototype system powered by the proposed patient similarity methods, and evaluate the proposed methods using real EHR data against several baselines.

[1]  Rong Jin,et al.  Distance Metric Learning: A Comprehensive Survey , 2006 .

[2]  Anthony Widjaja,et al.  Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond , 2003, IEEE Transactions on Neural Networks.

[3]  Gunnar Rätsch,et al.  Large Scale Multiple Kernel Learning , 2006, J. Mach. Learn. Res..

[4]  Kun Zhou,et al.  Locality Sensitive Discriminant Analysis , 2007, IJCAI.

[5]  Geoffrey E. Hinton,et al.  Stochastic Neighbor Embedding , 2002, NIPS.

[6]  D. Bates,et al.  Using Diagnoses to Describe Populations and Predict Costs , 2000, Health care financing review.

[7]  Yehuda Lindell,et al.  Privacy Preserving Data Mining , 2002, Journal of Cryptology.

[8]  Philip S. Yu,et al.  Privacy-Preserving Data Mining - Models and Algorithms , 2008, Advances in Database Systems.

[9]  Michael I. Jordan,et al.  Multiple kernel learning, conic duality, and the SMO algorithm , 2004, ICML.

[10]  Jimeng Sun,et al.  Integrating Distance Metrics Learned from Multiple Experts and its Application in Inter-Patient Similarity Assessment , 2011, SDM.

[11]  Geoffrey E. Hinton,et al.  Neighbourhood Components Analysis , 2004, NIPS.

[12]  Heng Tao Shen,et al.  Principal Component Analysis , 2009, Encyclopedia of Biometrics.

[13]  Michael I. Jordan,et al.  Distance Metric Learning with Application to Clustering with Side-Information , 2002, NIPS.

[14]  Fei Wang,et al.  Two Heads Better Than One: Metric+Active Learning and its Applications for IT Service Classification , 2009, 2009 Ninth IEEE International Conference on Data Mining.

[15]  Chris H. Q. Ding,et al.  Spectral Relaxation for K-means Clustering , 2001, NIPS.

[16]  Shigeo Abe DrEng Pattern Classification , 2001, Springer London.

[17]  Jimeng Sun,et al.  Localized Supervised Metric Learning on Temporal Physiological Data , 2010, 2010 20th International Conference on Pattern Recognition.

[18]  Chris Clifton,et al.  Privacy Preserving Data Mining (Advances in Information Security) , 2005 .

[19]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.

[20]  Arthur E. Hoerl,et al.  Ridge Regression: Biased Estimation for Nonorthogonal Problems , 2000, Technometrics.

[21]  Kilian Q. Weinberger,et al.  Distance Metric Learning for Large Margin Nearest Neighbor Classification , 2005, NIPS.

[22]  Jieping Ye,et al.  Computational and Theoretical Analysis of Null Space and Orthogonal Linear Discriminant Analysis , 2006, J. Mach. Learn. Res..

[23]  Yoram Singer,et al.  Efficient projections onto the l1-ball for learning in high dimensions , 2008, ICML '08.

[24]  Vladimir Vapnik,et al.  The Nature of Statistical Learning , 1995 .

[25]  Vladimir N. Vapnik,et al.  The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.

[26]  Nello Cristianini,et al.  Learning the Kernel Matrix with Semidefinite Programming , 2002, J. Mach. Learn. Res..

[27]  David G. Stork,et al.  Pattern Classification (2nd ed.) , 1999 .

[28]  Jun Liu,et al.  Efficient Euclidean projections in linear time , 2009, ICML '09.

[29]  Fei Wang Semisupervised Metric Learning by Maximizing Constraint Margin. , 2011 .

[30]  V. N. Bogaevski,et al.  Matrix Perturbation Theory , 1991 .

[31]  Fei Wang,et al.  iMet: Interactive Metric Learning in Healthcare Applications , 2011, SDM.

[32]  J. Friedman Regularized Discriminant Analysis , 1989 .

[33]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.