Ensemble Hidden Markov Models with Extended Observation Densities for Biosignal Analysis

Hidden Markov Models (HMM) have proven to be very useful in a variety of biomedical applications. The most established method for estimating HMM parameters is the maximum likelihood method which has shortcomings, such as repeated estimation and penalisation of the likelihood score, that are well known. This paper describes a variational learning approach to try and improve on the maximum-likelihood estimators. Emphasis lies on the fact that for HMMs with observation models that are from the exponential family of distributions, all HMM parameters and hidden state variables can be derived from a single loss function, namely the Kullback-Leibler divergence. Practical issues, such as model initialisation and choice of model order, are described. The paper concludes with application of three types of observation model HMMs to a variety of biomedical data, such as EEG and ECG, from different physiological experiments and conditions.

[1]  Jerry Nedelman,et al.  Book review: “Bayesian Data Analysis,” Second Edition by A. Gelman, J.B. Carlin, H.S. Stern, and D.B. Rubin Chapman & Hall/CRC, 2004 , 2005, Comput. Stat..

[2]  David Heckerman,et al.  A Tutorial on Learning with Bayesian Networks , 1998, Learning in Graphical Models.

[3]  Tommi S. Jaakkola,et al.  Tutorial on variational approximation methods , 2000 .

[4]  E. Wolpert A Manual of Standardized Terminology, Techniques and Scoring System for Sleep Stages of Human Subjects. , 1969 .

[5]  I. Rezek,et al.  Stochastic complexity measures for physiological signal analysis , 1998, IEEE Transactions on Biomedical Engineering.

[6]  Volker Tresp,et al.  Model-independent mean-field theory as a local method for approximate propagation of information. , 1999, Network.

[7]  Lawrence R. Rabiner,et al.  A tutorial on hidden Markov models and selected applications in speech recognition , 1989, Proc. IEEE.

[8]  Richard M. Everson,et al.  Hidden Markov Independent Component Analysis , 2000 .

[9]  David Heckerman,et al.  Learning With Bayesian Networks (Abstract) , 1995, ICML.

[10]  A. Rechtschaffen,et al.  A manual of standardized terminology, technique and scoring system for sleep stages of human subjects , 1968 .

[11]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[12]  David B. Dunson,et al.  Bayesian Data Analysis , 2010 .

[13]  Otto Rompelman,et al.  The Measurement of Heart Rate Variability Spectra with the Help of a Personal Computer , 1982, IEEE Transactions on Biomedical Engineering.

[14]  William D. Penny,et al.  Dynamic Models for Nonstationary Signal Segmentation , 1999, Comput. Biomed. Res..

[15]  Geoffrey E. Hinton,et al.  SMEM Algorithm for Mixture Models , 1998, Neural Computation.

[16]  Michael I. Jordan,et al.  Improving the Mean Field Approximation Via the Use of Mixture Distributions , 1999, Learning in Graphical Models.

[17]  A. Rukhin Matrix Variate Distributions , 1999, The Multivariate Normal Distribution.

[18]  C. Robert,et al.  Bayesian inference in hidden Markov models through the reversible jump Markov chain Monte Carlo method , 2000 .

[19]  B. Kemp,et al.  Model-based monitoring of human sleep stages , 1987 .

[20]  Michael I. Jordan,et al.  An Introduction to Variational Methods for Graphical Models , 1999, Machine-mediated learning.

[21]  Christa Neuper,et al.  Hidden Markov models for online classification of single trial EEG data , 2001, Pattern Recognit. Lett..

[22]  Stephen J. Roberts,et al.  Bayesian time series classification , 2001, NIPS.

[23]  Georg Dorffner,et al.  An automatic, continuous and probabilistic sleep stager based on a hidden markov model , 2002, Appl. Artif. Intell..

[24]  Steve Young,et al.  The HTK book , 1995 .