Leveraging Domain-expert Knowledge, Boosting and Deep Learning for Identification of Rare and Complex States

Identification of rare states in complex dynamical systems remains challenging due to data incompleteness and limited success in adopting standard machine learning algorithms for such problems. Although alternatives such as single-example learning (SEL) can learn novel classes from just a few training examples using similarities to reference classes for which large data sets are available, SEL performance quickly deteriorates with decreasing number of available reference classes. Previously, we have shown that with availability of just two broad classes, vector of boosting-ensemble components, discovered in supervised classification learning, provides universal state representation. Analytical models and other domain-expert knowledge can be directly incorporated within such ensemble decomposition learning (EDL) framework reducing requirements for data size. Alternatively, auto-encoders (AE), components of deep learning (DL) frameworks for automatic feature selection and dimensionality reduction, offer unlimited flexibility in discovering robust latent representations. However, AEs require large data sets and unlike boosting are less effective in utilization of existing models. We argue that using EDL vectors as inputs in standard AE architectures allows synergetic combination of the two approaches and discovery of robust state representations even with limited data. The framework is illustrated in the context of personalized medicine application using cardio data from www.physionet.org.

[1]  Jian Yang,et al.  Boosted Convolutional Neural Networks , 2016, BMVC.

[2]  Florian Metze,et al.  Extracting deep bottleneck features using stacked auto-encoders , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.

[3]  Shimon Edelman,et al.  Representation and recognition in vision , 1999 .

[4]  Fabrizio Lillo,et al.  Correlation, Hierarchies, and Networks in Financial Markets , 2008, 0809.4615.

[5]  Robert E. Schapire,et al.  Design and analysis of efficient learning algorithms , 1992, ACM Doctoral dissertation award ; 1991.

[6]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[7]  Valeriy Gavrishchaka,et al.  Multi-complexity ensemble measures for gait time series analysis: application to diagnostics, monitoring and biometrics. , 2015, Advances in experimental medicine and biology.

[8]  Geoffrey E. Hinton,et al.  Deep Learning , 2015, Nature.

[9]  Yan Liu,et al.  Distilling Knowledge from Deep Networks with Applications to Healthcare Domain , 2015, ArXiv.

[10]  Shai Shalev-Shwartz,et al.  SelfieBoost: A Boosting Algorithm for Deep Learning , 2014, ArXiv.

[11]  Pietro Perona,et al.  A Bayesian approach to unsupervised one-shot learning of object categories , 2003, Proceedings Ninth IEEE International Conference on Computer Vision.

[12]  Ronald M. Summers,et al.  Deep Convolutional Neural Networks for Computer-Aided Detection: CNN Architectures, Dataset Characteristics and Transfer Learning , 2016, IEEE Transactions on Medical Imaging.

[13]  Yoshua Bengio,et al.  Boosting Neural Networks , 2000, Neural Computation.

[14]  A. Malliani,et al.  Heart rate variability. Standards of measurement, physiological interpretation, and clinical use , 1996 .

[15]  Dong Yu,et al.  Deep Learning: Methods and Applications , 2014, Found. Trends Signal Process..

[16]  Jacques Bélair,et al.  Dynamical disease : mathematical analysis of human illness , 1995 .

[17]  Valeriy V. Gavrishchaka,et al.  Ensemble Decomposition Learning for Optimal Utilization of Implicitly Encoded Knowledge in Biomedical Applications , 2011 .

[18]  Ghassan Hamarneh,et al.  Pareto-optimal multi-objective dimensionality reduction deep auto-encoder for mammography classification , 2017, Comput. Methods Programs Biomed..

[19]  J. Friedman Special Invited Paper-Additive logistic regression: A statistical view of boosting , 2000 .

[20]  Amanda J. C. Sharkey,et al.  Combining Artificial Neural Nets: Ensemble and Modular Multi-Net Systems , 1999 .

[21]  Paul A. Viola,et al.  Learning from one example through shared densities on transforms , 2000, Proceedings IEEE Conference on Computer Vision and Pattern Recognition. CVPR 2000 (Cat. No.PR00662).

[22]  Valeriy V. Gavrishchaka,et al.  Boosting-Based Frameworks in Financial Modeling: Application to Symbolic Volatility Forecasting , 2006 .

[23]  Bo Yang,et al.  Towards K-means-friendly Spaces: Simultaneous Deep Learning and Clustering , 2016, ICML.

[24]  Vipin Kumar,et al.  Evaluating boosting algorithms to classify rare classes: comparison and improvements , 2001, Proceedings 2001 IEEE International Conference on Data Mining.

[25]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[26]  J. Friedman Greedy function approximation: A gradient boosting machine. , 2001 .

[27]  Jasper Snoek,et al.  Nonparametric guidance of autoencoder representations using label information , 2012, J. Mach. Learn. Res..

[28]  Olga V. Senyukova,et al.  Generic Ensemble-Based Representation of Global Cardiovascular Dynamics for Personalized Treatment Discovery and Optimization , 2016, ICCCI.

[29]  K. Kaski,et al.  Dynamics of market correlations: taxonomy and portfolio analysis. , 2003, Physical review. E, Statistical, nonlinear, and soft matter physics.

[30]  T. Breuel A Bayesian Approach to Learning Single View Generalization in 3D Object Recognition , 2003 .

[31]  Michael Fink,et al.  Object Classification from a Single Example Utilizing Class Relevance Metrics , 2004, NIPS.

[32]  Marios Anthimopoulos,et al.  Multi-source Transfer Learning with Convolutional Neural Networks for Lung Pattern Analysis , 2016, IEEE journal of biomedical and health informatics.

[33]  Shimon Ullman,et al.  Single-example Learning of Novel Classes using Representation by Similarity , 2005, BMVC.

[34]  Colin Raffel,et al.  Realistic Evaluation of Semi-Supervised Learning Algorithms , 2018, ICLR.