Classification of sparsely and irregularly sampled time series: A learning in model space approach

Classification of sparsely and irregularly sampled time series data is a challenging machine learning task. To tackle this problem, we present a learning in model space framework in which time-continuous dynamical system models are first inferred from individual time series and then the inferred models are used to represent these time series for the classification task. In contrast to the existing approaches using model point estimates to represent individual time series, we further employ posterior distributions over models, thus taking into account in a principled manner the uncertainty around the inferred model due to observation noise and data sparsity. Finally, we present a distributional classifier for classifying the posterior distributions. We evaluate the framework on a biological pathway model. In particular, we investigate the classification performance in the cases where model uncertainties in the training and test phases do not match.

[1]  Andreas Ruttor,et al.  Approximate Gaussian process inference for the drift function in stochastic differential equations , 2013, NIPS.

[2]  Krasimira Tsaneva-Atanasova,et al.  Decoding GnRH neurohormone pulse frequency by convergent signalling modules , 2012, Journal of the Royal Society Interface.

[3]  Joachim M. Buhmann,et al.  Generative Embedding for Model-Based Classification of fMRI Data , 2011, PLoS Comput. Biol..

[4]  Charu C. Aggarwal,et al.  On effective classification of strings with wavelets , 2002, KDD.

[5]  Eamonn J. Keogh,et al.  Scaling up dynamic time warping for datamining applications , 2000, KDD '00.

[6]  Alexander J. Smola,et al.  Binet-Cauchy Kernels on Dynamical Systems and its Application to the Analysis of Dynamic Scenes , 2007, International Journal of Computer Vision.

[7]  T. Warren Liao,et al.  Clustering of time series data - a survey , 2005, Pattern Recognit..

[8]  Huanhuan Chen,et al.  Model Metric Co-Learning for Time Series Classification , 2015, IJCAI.

[9]  Nuno Vasconcelos,et al.  A Kullback-Leibler Divergence Based Kernel for SVM Classification in Multimedia Applications , 2003, NIPS.

[10]  Tony Jebara,et al.  Probability Product Kernels , 2004, J. Mach. Learn. Res..

[11]  Alexander J. Smola,et al.  Binet-Cauchy Kernels , 2004, NIPS.

[12]  Harry van Zanten,et al.  Reversible jump MCMC for nonparametric drift estimation for diffusion processes , 2012, Comput. Stat. Data Anal..

[13]  Nuno Vasconcelos,et al.  Probabilistic kernels for the classification of auto-regressive visual processes , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[14]  Stefano Soatto,et al.  Classification and Recognition of Dynamical Models: The Role of Phase, Independent Components, Kernels and Optimal Transport , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[15]  Mark A. Girolami,et al.  Bayesian inference for differential equations , 2008, Theor. Comput. Sci..

[16]  Mohammed J. Zaki,et al.  Mining features for sequence classification , 1999, KDD '99.

[17]  Le Song,et al.  A Hilbert Space Embedding for Distributions , 2007, Discovery Science.

[18]  Bernhard Schölkopf,et al.  Learning from Distributions via Support Measure Machines , 2012, NIPS.

[19]  Jian Pei,et al.  A brief survey on sequence classification , 2010, SKDD.