Automatic Backward Filtering Forward Guiding for Markov processes and graphical models

We incorporate discrete and continuous time Markov processes as building blocks into probabilistic graphical models. Observations are represented by leaf vertices. We introduce the automatic Backward Filtering Forward Guiding (BFFG) paradigm (Mider et. al, 2020) for programmable inference on latent states and model parameters. Our starting point is a generative model, a forward description of the probabilistic process dynamics. We backpropagate the information provided by observations through the model to transform the generative (forward) model into a pre-conditional model guided by the data. It approximates the actual conditional model with known likelihood-ratio between the two. The backward filter and the forward change of measure are suitable to be incorporated into a probabilistic programming context because they can be formulated as a set of transformation rules. The guided generative model can be incorporated in different approaches to efficiently sample latent states and parameters conditional on observations. We show applicability in a variety of settings, including Markov chains with discrete state space, interacting particle systems, state space models, branching diffusions and Gamma processes.

[1]  Arnaud Doucet,et al.  Sequential Monte Carlo Methods , 2006, Handbook of Graphical Models.

[2]  Jonathan Law,et al.  Functional probabilistic programming for scalable Bayesian modelling , 2019, 1908.02062.

[3]  Tobias Fritz A synthetic approach to Markov kernels, conditional independence, and theorems on sufficient statistics , 2019, ArXiv.

[4]  P. Moral,et al.  Tree based functional expansions for Feynman–Kac particle models , 2009, 0906.4249.

[5]  M. J. Bayarri,et al.  Non-Centered Parameterisations for Hierarchical Models and Data Augmentation , 2003 .

[7]  Moritz Schauer,et al.  Continuous-discrete smoothing of diffusions , 2017, Electronic Journal of Statistics.

[8]  V. Spokoiny,et al.  Forward and reverse representations for Markov chains ∗ , 2006 .

[9]  B. Khattatov,et al.  Data assimilation : making sense of observations , 2010 .

[10]  Alexandros Beskos,et al.  Online Smoothing for Diffusion Processes Observed with Noise , 2020 .

[11]  Bayesian inference for discretely observed diffusion processes , 2015 .

[12]  Max Welling,et al.  Auto-Encoding Variational Bayes , 2013, ICLR.

[13]  Judea Pearl,et al.  Probabilistic reasoning in intelligent systems - networks of plausible inference , 1991, Morgan Kaufmann series in representation and reasoning.

[14]  Paul Fearnhead,et al.  Inference for reaction networks using the linear noise approximation , 2012, Biometrics.

[15]  B. Delyon,et al.  Simulation of conditioned diffusion and application to parameter estimation , 2006 .

[16]  Prakash Panangaden,et al.  The Category of Markov Kernels , 1998, PROBMIV.

[17]  P. Selinger A Survey of Graphical Languages for Monoidal Categories , 2009, 0908.3347.

[18]  Stefan Grosskinsky Warwick,et al.  Interacting Particle Systems , 2016 .

[19]  David Duvenaud,et al.  Neural Ordinary Differential Equations , 2018, NeurIPS.

[20]  Fredrik Bagge Carlson MonteCarloMeasurements.jl: Nonlinear Propagation of Arbitrary Multivariate Distributions by means of Method Overloading , 2020, ArXiv.

[21]  Zbigniew Palmowski,et al.  The technique of the exponential change of measure for Markov processes , 2002 .

[22]  Darryl D. Holm Variational principles for stochastic fluid dynamics , 2014, Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences.

[23]  Wojciech Pieczynski,et al.  Kalman filtering in pairwise Markov trees , 2006, Signal Process..

[24]  L. Rogers,et al.  Diffusions, Markov processes, and martingales , 1979 .

[25]  Harry van Zanten,et al.  Guided proposals for simulating multi-dimensional diffusion bridges , 2013, 1311.3606.

[26]  A. Hasman,et al.  Probabilistic reasoning in intelligent systems: Networks of plausible inference , 1991 .

[27]  Thomas G. Kurtz,et al.  Stochastic Analysis of Biochemical Systems , 2015 .

[28]  Stanley Osher,et al.  ResNets Ensemble via the Feynman-Kac Formalism to Improve Natural and Robust Accuracies , 2018, NeurIPS.

[29]  R. Kohn,et al.  On Gibbs sampling for state space models , 1994 .

[30]  Shakir Mohamed,et al.  Variational Inference with Normalizing Flows , 2015, ICML.

[31]  Michael Innes,et al.  Don't Unroll Adjoint: Differentiating SSA-Form Programs , 2018, ArXiv.

[32]  John Schoenmakers,et al.  Transition density estimation for stochastic differential equations via forward-reverse representations , 2004 .

[33]  K. C. Chou,et al.  Multiscale recursive estimation, data fusion, and regularization , 1994, IEEE Trans. Autom. Control..

[34]  Mitchell Riley Categories of Optics , 2018 .

[35]  F. Meulen,et al.  Bayesian estimation of discretely observed multi-dimensional diffusion processes using guided proposals , 2014, Electronic Journal of Statistics.

[36]  C. Andrieu,et al.  The pseudo-marginal approach for efficient Monte Carlo computations , 2009, 0903.5480.

[37]  Ralph S. Silva,et al.  On Some Properties of Markov Chain Monte Carlo Simulation Methods Based on the Particle Filter , 2012 .

[38]  Ali Ramadhan,et al.  Universal Differential Equations for Scientific Machine Learning , 2020, ArXiv.

[39]  J. Bierkens,et al.  Simulation of elliptic and hypo-elliptic conditional diffusions , 2018, Advances in Applied Probability.

[40]  Donald L. Iglehart,et al.  Importance sampling for stochastic simulations , 1989 .

[41]  Francois Desbouvries,et al.  Exact Bayesian Prediction in a Class of Markov-switching Models , 2012 .

[42]  Ivan Kobyzev,et al.  Normalizing Flows: An Introduction and Review of Current Methods , 2020, IEEE Transactions on Pattern Analysis and Machine Intelligence.