Compressed Monte Carlo with application in particle filtering

Abstract Bayesian models have become very popular over the last years in several fields such as signal processing, statistics, and machine learning. Bayesian inference requires the approximation of complicated integrals involving posterior distributions. For this purpose, Monte Carlo (MC) methods, such as Markov Chain Monte Carlo and importance sampling algorithms, are often employed. In this work, we introduce the theory and practice of a Compressed MC (C-MC) scheme to compress the statistical information contained in a set of random samples. In its basic version, C-MC is strictly related to the stratification technique, a well-known method used for variance reduction purposes. Deterministic C-MC schemes are also presented, which provide very good performance. The compression problem is strictly related to the moment matching approach applied in different filtering techniques, usually called as Gaussian quadrature rules or sigma-point methods. C-MC can be employed in a distributed Bayesian inference framework when cheap and fast communications with a central processor are required. Furthermore, C-MC is useful within particle filtering and adaptive IS algorithms, as shown by three novel schemes introduced in this work. Six numerical results confirm the benefits of the introduced schemes, outperforming the corresponding benchmark methods. A related code is also provided. (The code is provided at  http://www.lucamartino.altervista.org/CMC_CODE_pub_EX1.zip )

[1]  O. Lahav,et al.  exofit: orbital parameters of extrasolar planets from radial velocities , 2008, 0805.3532.

[2]  A.S. Willsky,et al.  Distributed fusion in sensor networks , 2006, IEEE Signal Processing Magazine.

[3]  Michael G. Rabbat,et al.  Particle weight approximation with clustering for gossip-based distributed particle filters , 2015, 2015 IEEE 6th International Workshop on Computational Advances in Multi-Sensor Adaptive Processing (CAMSAP).

[4]  John W. Fisher,et al.  Nonparametric belief propagation for self-localization of sensor networks , 2005, IEEE Journal on Selected Areas in Communications.

[5]  Petar M. Djuric,et al.  Consensus-based Distributed Particle Filtering With Distributed Proposal Adaptation , 2014, IEEE Transactions on Signal Processing.

[6]  Jukka Corander,et al.  Layered adaptive importance sampling , 2015, Statistics and Computing.

[7]  William T. Freeman,et al.  Efficient Multiscale Sampling from Products of Gaussian Mixtures , 2003, NIPS.

[8]  Mark Coates,et al.  Asynchronous distributed particle filter via decentralized evaluation of Gaussian products , 2010, 2010 13th International Conference on Information Fusion.

[9]  Harald Niederreiter,et al.  Random number generation and Quasi-Monte Carlo methods , 1992, CBMS-NSF regional conference series in applied mathematics.

[10]  Petar M. Djuric,et al.  Gaussian sum particle filtering , 2003, IEEE Trans. Signal Process..

[11]  L. Pronzato Minimax and maximin space-filling designs: some properties and methods for construction , 2017 .

[12]  J. Marin,et al.  Population Monte Carlo , 2004 .

[13]  Yuanxin Wu,et al.  A Numerical-Integration Perspective on Gaussian Filters , 2006, IEEE Transactions on Signal Processing.

[14]  Luca Martino,et al.  Group Importance Sampling for Particle Filtering and MCMC , 2017, Digit. Signal Process..

[15]  Petar M. Djuric,et al.  Resampling algorithms and architectures for distributed particle filters , 2005, IEEE Transactions on Signal Processing.

[16]  V. Roshan Joseph,et al.  Support points , 2016, The Annals of Statistics.

[17]  Luca Martino,et al.  Different acceptance functions for Multiple Try Metropolis schemes , 2012 .

[18]  W. Marsden I and J , 2012 .

[19]  Amir Asif,et al.  Diffusive particle filtering for distributed multisensor estimation , 2016, 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[20]  Ingmar Schuster,et al.  Markov Chain Importance Sampling—A Highly Efficient Estimator for MCMC , 2018, J. Comput. Graph. Stat..

[21]  Joaquín Míguez,et al.  A proof of uniform convergence over time for a distributed particle filter , 2015, Signal Process..

[22]  Petar M. Djuric,et al.  Gaussian particle filtering , 2003, IEEE Trans. Signal Process..

[23]  Jesse Read,et al.  A distributed particle filter for nonlinear tracking in wireless sensor networks , 2014, Signal Process..

[24]  Mónica F. Bugallo,et al.  Multiple Particle Filtering , 2007, 2007 IEEE International Conference on Acoustics, Speech and Signal Processing - ICASSP '07.

[25]  Simo Särkkä,et al.  Bayesian Filtering and Smoothing , 2013, Institute of Mathematical Statistics textbooks.

[26]  James R. Wilson Variance Reduction Techniques for Digital Simulation , 1984 .

[27]  Simo Srkk,et al.  Bayesian Filtering and Smoothing , 2013 .

[28]  Luca Martino,et al.  Cooperative parallel particle filters for online model selection and applications to urban mobility , 2015, Digit. Signal Process..

[29]  Jean-Michel Marin,et al.  Adaptive importance sampling in general mixture classes , 2007, Stat. Comput..

[30]  Lester W. Mackey,et al.  Stein Points , 2018, ICML.

[31]  P. Fearnhead Using Random Quasi-Monte-Carlo Within Particle Filters, With Application to Financial Time Series , 2005 .

[32]  Nando de Freitas,et al.  Fast particle smoothing: if I had a million particles , 2006, ICML.

[33]  P. Alam ‘N’ , 2021, Composites Engineering: An A–Z Guide.

[34]  Petar M. Djuric,et al.  Adaptive Importance Sampling: The past, the present, and the future , 2017, IEEE Signal Processing Magazine.

[35]  Luca Martino,et al.  Weighting a resampled particle in Sequential Monte Carlo , 2016, 2016 IEEE Statistical Signal Processing Workshop (SSP).

[36]  Pierre L'Ecuyer,et al.  Efficiency improvement and variance reduction , 1994, Proceedings of Winter Simulation Conference.

[37]  V. Roshan Joseph,et al.  Projected support points: a new method for high-dimensional data reduction. , 2017, 1708.06897.

[38]  Tiancheng Li,et al.  Deterministic resampling: Unbiased sampling to avoid sample impoverishment in particle filters , 2012, Signal Process..

[39]  H. Vincent Poor,et al.  Distributed learning in wireless sensor networks , 2005, IEEE Signal Processing Magazine.

[40]  Eric Moulines,et al.  On parallel implementation of sequential Monte Carlo methods: the island particle model , 2013, Stat. Comput..

[41]  Edward I. George,et al.  Bayes and big data: the consensus Monte Carlo algorithm , 2016, Big Data and Information Theory.

[42]  Fredrik Lindsten,et al.  Sequential Kernel Herding: Frank-Wolfe Optimization for Particle Filtering , 2015, AISTATS.

[43]  Jun S. Liu,et al.  Monte Carlo strategies in scientific computing , 2001 .

[44]  Petar M. Djuric,et al.  Resampling Methods for Particle Filtering: Classification, implementation, and strategies , 2015, IEEE Signal Processing Magazine.

[45]  David Duvenaud,et al.  Optimally-Weighted Herding is Bayesian Quadrature , 2012, UAI.

[46]  O. Cappé,et al.  Population Monte Carlo , 2004 .

[47]  Jeffrey K. Uhlmann,et al.  Unscented filtering and nonlinear estimation , 2004, Proceedings of the IEEE.

[48]  S. Haykin,et al.  Cubature Kalman Filters , 2009, IEEE Transactions on Automatic Control.

[49]  Stergios I. Roumeliotis,et al.  Set-Membership Constrained Particle Filter: Distributed Adaptation for Sensor Networks , 2011, IEEE Transactions on Signal Processing.

[50]  Alexander J. Smola,et al.  Super-Samples from Kernel Herding , 2010, UAI.

[51]  Christian P. Robert,et al.  Monte Carlo Statistical Methods , 2005, Springer Texts in Statistics.