Learning Stationary Time Series using Gaussian Processes with Nonparametric Kernels

We introduce the Gaussian Process Convolution Model (GPCM), a two-stage non-parametric generative procedure to model stationary signals as the convolution between a continuous-time white-noise process and a continuous-time linear filter drawn from Gaussian process. The GPCM is a continuous-time nonparametric-window moving average process and, conditionally, is itself a Gaussian process with a nonparametric kernel defined in a probabilistic fashion. The generative model can be equivalently considered in the frequency domain, where the power spectral density of the signal is specified using a Gaussian process. One of the main contributions of the paper is to develop a novel variational free-energy approach based on inter-domain inducing variables that efficiently learns the continuous-time linear filter and infers the driving white-noise process. In turn, this scheme provides closed-form probabilistic estimates of the covariance kernel and the noise-free signal both in denoising and prediction scenarios. Additionally, the variational inference procedure provides closed-form expressions for the approximate posterior of the spectral density given the observed data, leading to new Bayesian nonparametric approaches to spectrum estimation. The proposed GPCM is validated using synthetic and real-world signals.

[1]  David J. C. MacKay,et al.  Information Theory, Inference, and Learning Algorithms , 2004, IEEE Transactions on Information Theory.

[2]  Richard E. Turner,et al.  Time-Frequency Analysis as Probabilistic Inference , 2014, IEEE Transactions on Signal Processing.

[3]  T. Hughes,et al.  Signals and systems , 2006, Genome Biology.

[4]  Florian Nadel,et al.  Stochastic Processes And Filtering Theory , 2016 .

[5]  Richard E. Turner,et al.  Tree-structured Gaussian Process Approximations , 2014, NIPS.

[6]  Andrew Gordon Wilson,et al.  Gaussian Process Kernels for Pattern Discovery and Extrapolation , 2013, ICML.

[7]  James Hensman,et al.  On Sparse Variational Methods and the Kullback-Leibler Divergence between Stochastic Processes , 2015, AISTATS.

[8]  Sun-Yuan Kung,et al.  Multikernel Least Mean Square Algorithm , 2014, IEEE Transactions on Neural Networks and Learning Systems.

[9]  Donald B. Percival,et al.  Spectral Analysis for Physical Applications , 1993 .

[10]  Richard E. Turner,et al.  Two problems with variational expectation maximisation for time-series models , 2011 .

[11]  Dan Cornford,et al.  Gaussian Process Approximations of Stochastic Differential Equations , 2007, Gaussian Processes in Practice.

[12]  Carl E. Rasmussen,et al.  Additive Gaussian Processes , 2011, NIPS.

[13]  Carl E. Rasmussen,et al.  Gaussian processes for machine learning , 2005, Adaptive computation and machine learning.

[14]  Richard E. Turner Statistical models for natural sounds , 2010 .

[15]  Ethem Alpaydin,et al.  Multiple Kernel Learning Algorithms , 2011, J. Mach. Learn. Res..

[16]  Yuan Qi,et al.  Bayesian spectrum estimation of unevenly sampled nonstationary data , 2002, 2002 IEEE International Conference on Acoustics, Speech, and Signal Processing.

[17]  Michalis K. Titsias,et al.  Variational Learning of Inducing Variables in Sparse Gaussian Processes , 2009, AISTATS.

[18]  Thang D. Bui,et al.  Design of Covariance Functions using Inter-Domain Inducing Variables , 2015 .

[19]  Anahí Gallardo Velázquez,et al.  Conference , 1969, Journal of Neuroscience Methods.

[20]  Anthony Widjaja,et al.  Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond , 2003, IEEE Transactions on Neural Networks.

[21]  S. Shreve,et al.  Stochastic differential equations , 1955, Mathematical Proceedings of the Cambridge Philosophical Society.

[22]  Aníbal R. Figueiras-Vidal,et al.  Inter-domain Gaussian Processes for Sparse Inference using Inducing Features , 2009, NIPS.

[23]  B. Øksendal Stochastic Differential Equations , 1985 .

[24]  Yoshua. Bengio,et al.  Learning Deep Architectures for AI , 2007, Found. Trends Mach. Learn..

[25]  Michael I. Jordan,et al.  Learning with Mixtures of Trees , 2001, J. Mach. Learn. Res..

[26]  Neil D. Lawrence,et al.  Bayesian Gaussian Process Latent Variable Model , 2010, AISTATS.

[27]  Iain Murray Introduction To Gaussian Processes , 2008 .

[28]  Alexander J. Smola,et al.  Neural Information Processing Systems , 1997, NIPS 1997.

[29]  Joshua B. Tenenbaum,et al.  Structure Discovery in Nonparametric Regression through Compositional Kernel Search , 2013, ICML.

[30]  Stephen F. Gull,et al.  Developments in Maximum Entropy Data Analysis , 1989 .