Bayesian imaging using Plug & Play priors: when Langevin meets Tweedie

Since the seminal work of Venkatakrishnan et al. [89] in 2013, Plug & Play (PnP) methods have become ubiquitous in Bayesian imaging. These methods derive Minimum Mean Square Error (MMSE) or Maximum A Posteriori (MAP) estimators for inverse problems in imaging by combining an explicit likelihood function with a prior that is implicitly defined by an image denoising algorithm. The PnP algorithms proposed in the literature mainly differ in the iterative schemes they use for optimisation or for sampling. In the case of optimisation schemes, some recent works guarantee the convergence to a fixed point, albeit not necessarily a MAP estimate. In the case of sampling schemes, to the best of our knowledge, there is no known proof of convergence. There also remain important open questions regarding whether the underlying Bayesian models and estimators are well defined, well-posed, and have the basic regularity properties required to support these numerical schemes. To address these limitations, this paper develops theory, methods, and provably convergent algorithms for performing Bayesian inference with PnP priors. We introduce two algorithms: 1) PnPULA (Plug & Play Unadjusted Langevin Algorithm) for Monte Carlo sampling and MMSE inference; and 2) PnP-SGD (Plug & Play Stochastic Gradient Descent) for MAP inference. Using recent results on the quantitative convergence of Markov chains, we establish detailed convergence guarantees for these two algorithms under realistic assumptions on the denoising operators used, with special attention to denoisers based on deep neural networks. We also show that these algorithms approximately target a decision-theoretically optimal Bayesian model that is well-posed. The proposed algorithms are demonstrated on several canonical problems such as image deblurring, inpainting, and denoising, where they are used for point estimation as well as for uncertainty visualisation and quantification. ∗VDB was partially supported by EPSRC grant EP/R034710/1. RL was partially supported by grants from Région Ile-De-France. AD acknowledges support of the Lagrange Mathematical and Computing Research Center. MP was partially supported by EPSRC grant EP/T007346/1. JD and AA acknowledge support from the French Research Agency through the PostProdLEAP project (ANR-19-CE23-0027-01). Computer experiments for this work ran on a Titan Xp GPU donated by NVIDIA, as well as on HPC resources from GENCI-IDRIS (Grant 2020-AD011011641). †These authors contributed equally ‡Université de Paris, MAP5 UMR 8145, F-75006 Paris, France §Department of Statistics University of Oxford 24-29 St Giles OX1 3LB, Oxford United Kingdom ¶Institut Universitaire de France (IUF) ‖Centre Borelli, UMR 9010, École Normale Supérieure Paris-Saclay ∗∗School of Mathematical and Computer Sciences, Heriot-Watt University & Maxwell Institute for Mathematical Sciences, Edinburgh, United Kingdom 1 ar X iv :2 10 3. 04 71 5v 4 [ st at .M E ] 1 9 M ar 2 02 1

[1]  Julie Delon,et al.  On Maximum a Posteriori Estimation with Plug & Play Priors and Stochastic Gradient Descent , 2022, Journal of Mathematical Imaging and Vision.

[2]  Michael Elad,et al.  Stochastic Image Denoising by Sampling from the Posterior Distribution , 2021, 2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW).

[3]  Gershon Wolansky,et al.  Optimal Transport , 2021 .

[4]  Jean-Christophe Pesquet,et al.  Learning Maximally Monotone Operators for Image Recovery , 2020, SIAM J. Imaging Sci..

[5]  Michael Elad,et al.  Regularization by Denoising via Fixed-Point Projection (RED-PRO) , 2020, SIAM J. Imaging Sci..

[6]  Eero P. Simoncelli,et al.  Stochastic Solutions for Linear Inverse Problems using the Prior Implicit in a Denoiser , 2021, NeurIPS.

[7]  Alain Durmus,et al.  Maximum Likelihood Estimation of Regularization Parameters in High-Dimensional Inverse Problems: An Empirical Bayesian Approach. Part II: Theoretical Analysis , 2020, SIAM J. Imaging Sci..

[8]  C. Schonlieb,et al.  Learned convex regularizers for inverse problems , 2020, ArXiv.

[9]  Eero P. Simoncelli,et al.  Solving Linear Inverse Problems Using the Prior Implicit in a Denoiser , 2020, ArXiv.

[10]  Pieter Abbeel,et al.  Denoising Diffusion Probabilistic Models , 2020, NeurIPS.

[11]  Brendt Wohlberg,et al.  Provable Convergence of Plug-and-Play Priors With MMSE Denoisers , 2020, IEEE Signal Processing Letters.

[12]  K. Kunisch,et al.  Total Deep Variation for Linear Inverse Problems , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[13]  Valentin De Bortoli,et al.  Maximum Likelihood Estimation of Regularization Parameters in High-Dimensional Inverse Problems: An Empirical Bayesian Approach Part I: Methodology and Experiments , 2019, SIAM J. Imaging Sci..

[14]  Charles A. Bouman,et al.  Plug-and-Play Methods for Magnetic Resonance Imaging: Using Denoisers for Image Recovery , 2019, IEEE Signal Processing Magazine.

[15]  Jonas Latz,et al.  On the Well-posedness of Bayesian Inverse Problems , 2019, SIAM/ASA J. Uncertain. Quantification.

[16]  Francesco Renna,et al.  On instabilities of deep learning in image reconstruction and the potential costs of AI , 2019, Proceedings of the National Academy of Sciences.

[17]  K. Zygalakis,et al.  Accelerating Proximal Markov Chain Monte Carlo by Using an Explicit Stabilized Method | SIAM Journal on Imaging Sciences | Vol. 13, No. 2 | Society for Industrial and Applied Mathematics , 2020 .

[18]  Yang Song,et al.  Generative Modeling by Estimating Gradients of the Data Distribution , 2019, NeurIPS.

[19]  Xiaoyong Shen,et al.  Dynamic Scene Deblurring With Parameter Selective Sharing and Nested Skip Connections , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[20]  Xiaohan Chen,et al.  Plug-and-Play Methods Provably Converge with Properly Trained Denoisers , 2019, ICML.

[21]  Simon R. Arridge,et al.  Solving inverse problems using data-driven models , 2019, Acta Numerica.

[22]  Rebecca Willett,et al.  Neumann Networks for Inverse Problems in Imaging , 2019, ArXiv.

[23]  Li Shen,et al.  A Sufficient Condition for Convergences of Adam and RMSProp , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[24]  Philip Schniter,et al.  Regularization by Denoising: Clarifications and New Interpretations , 2018, IEEE Transactions on Computational Imaging.

[25]  Audrey Repetti,et al.  Scalable Bayesian Uncertainty Quantification in Imaging Inverse Problems via Convex Optimization , 2018, SIAM J. Imaging Sci..

[26]  Raja Giryes,et al.  DeepISP: Toward Learning an End-to-End Image Processing Pipeline , 2018, IEEE Transactions on Image Processing.

[27]  Marcelo Pereyra,et al.  Revisiting Maximum-A-Posteriori Estimation in Log-Concave Models , 2016, SIAM J. Imaging Sci..

[28]  Yuxing Han,et al.  AGEM: Solving Linear Inverse Problems via Deep Priors and Sampling , 2019, NeurIPS.

[29]  José M. Bioucas-Dias,et al.  A Convergent Image Fusion Algorithm Using Scene-Adapted Gaussian-Mixture-Based Denoising , 2019, IEEE Transactions on Image Processing.

[30]  Charles Bouveyron,et al.  High-Dimensional Mixture Models for Unsupervised Image Denoising (HDMI) , 2018, SIAM J. Imaging Sci..

[31]  Le Lu,et al.  DeepLesion: automated mining of large-scale lesion annotations and universal lesion detection with deep learning , 2018, Journal of medical imaging.

[32]  Sundeep Rangan,et al.  Plug in estimation in high dimensional linear inverse problems a rigorous analysis , 2018, NeurIPS.

[33]  Yuichi Yoshida,et al.  Spectral Normalization for Generative Adversarial Networks , 2018, ICLR.

[34]  José M. Bioucas-Dias,et al.  Scene-Adapted Plug-and-Play Algorithm with Guaranteed Convergence: Applications to Data Fusion in Imaging , 2018, ArXiv.

[35]  Marcelo Pereyra,et al.  Uncertainty quantification for radio interferometric imaging: I. proximal MCMC methods , 2017, Monthly Notices of the Royal Astronomical Society.

[36]  Lei Zhang,et al.  FFDNet: Toward a Fast and Flexible Solution for CNN-Based Image Denoising , 2017, IEEE Transactions on Image Processing.

[37]  Michael I. Jordan,et al.  Underdamped Langevin MCMC: A non-asymptotic analysis , 2017, COLT.

[38]  Charles A. Bouman,et al.  Plug-and-Play Unplugged: Optimization Free Reconstruction using Consensus Equilibrium , 2017, SIAM J. Imaging Sci..

[39]  Eric Moulines,et al.  Efficient Bayesian Computation by Proximal Markov Chain Monte Carlo: When Langevin Meets Moreau , 2016, SIAM J. Imaging Sci..

[40]  Jorge Nocedal,et al.  Optimization Methods for Large-Scale Machine Learning , 2016, SIAM Rev..

[41]  Hongqiang Wang,et al.  BM3D vector approximate message passing for radar coded-aperture imaging , 2017, 2017 Progress in Electromagnetics Research Symposium - Fall (PIERS - FALL).

[42]  Hassan Mansour,et al.  A Plug-and-Play Priors Approach for Solving Nonlinear Imaging Inverse Problems , 2017, IEEE Signal Processing Letters.

[43]  Matthias Zwicker,et al.  Deep Mean-Shift Priors for Image Restoration , 2017, NIPS.

[44]  Gordon Wetzstein,et al.  Unrolled Optimization with Deep Priors , 2017, ArXiv.

[45]  Julie Delon,et al.  A Bayesian Hyperprior Approach for Joint Image Denoising and Interpolation, With an Application to HDR Imaging , 2017, IEEE Transactions on Computational Imaging.

[46]  Michael Möller,et al.  Learning Proximal Operators: Using Denoising Networks for Regularizing Inverse Imaging Problems , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[47]  Wangmeng Zuo,et al.  Learning Deep CNN Denoiser Prior for Image Restoration , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[48]  Alexandros G. Dimakis,et al.  Compressed Sensing using Generative Models , 2017, ICML.

[49]  Michael Elad,et al.  The Little Engine That Could: Regularization by Denoising (RED) , 2016, SIAM J. Imaging Sci..

[50]  Lei Zhang,et al.  Beyond a Gaussian Denoiser: Residual Learning of Deep CNN for Image Denoising , 2016, IEEE Transactions on Image Processing.

[51]  Stanley H. Chan,et al.  Plug-and-Play ADMM for Image Restoration: Fixed-Point Convergence and Applications , 2016, IEEE Transactions on Computational Imaging.

[52]  Yunjin Chen,et al.  Trainable Nonlinear Reaction Diffusion: A Flexible Framework for Fast and Effective Image Restoration , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[53]  Francis R. Bach,et al.  Breaking the Curse of Dimensionality with Convex Neural Networks , 2014, J. Mach. Learn. Res..

[54]  Frédo Durand,et al.  Deep joint demosaicking and denoising , 2016, ACM Trans. Graph..

[55]  Charles A. Bouman,et al.  Plug-and-Play Priors for Bright Field Electron Tomography and Sparse Interpolation , 2015, IEEE Transactions on Computational Imaging.

[56]  Alfred O. Hero,et al.  A Survey of Stochastic Simulation and Optimization Methods in Signal Processing , 2015, IEEE Journal of Selected Topics in Signal Processing.

[57]  Richard G. Baraniuk,et al.  From Denoising to Compressed Sensing , 2014, IEEE Transactions on Information Theory.

[58]  Marcelo Pereyra,et al.  Proximal Markov chain Monte Carlo algorithms , 2013, Statistics and Computing.

[59]  É. Moulines,et al.  Non-asymptotic convergence analysis for the Unadjusted Langevin Algorithm , 2015, 1507.05021.

[60]  José M. Bioucas-Dias,et al.  Collaborative sparse regression using spatially correlated supports - Application to hyperspectral unmixing , 2014, IEEE Transactions on Image Processing.

[61]  A. Dalalyan Theoretical guarantees for approximate sampling from smooth and log‐concave densities , 2014, 1412.7392.

[62]  Xiaoou Tang,et al.  Learning a Deep Convolutional Network for Image Super-Resolution , 2014, ECCV.

[63]  C. Holmes,et al.  Approximate Models and Robust Decisions , 2014, 1402.6118.

[64]  Tianqi Chen,et al.  Stochastic Gradient Hamiltonian Monte Carlo , 2014, ICML.

[65]  Yoshua Bengio,et al.  What regularized auto-encoders learn from the data-generating distribution , 2012, J. Mach. Learn. Res..

[66]  Lionel Moisan,et al.  Posterior Expectation of the Total Variation Model: Properties and Experiments , 2013, SIAM J. Imaging Sci..

[67]  Brendt Wohlberg,et al.  Plug-and-Play priors for model based reconstruction , 2013, 2013 IEEE Global Conference on Signal and Information Processing.

[68]  Jean-Michel Morel,et al.  A Nonlocal Bayesian Image Denoising Algorithm , 2013, SIAM J. Imaging Sci..

[69]  Peyman Milanfar,et al.  Symmetrizing Smoothing Filters , 2013, SIAM J. Imaging Sci..

[70]  Adel Javanmard,et al.  State Evolution for General Approximate Message Passing Algorithms, with Applications to Spatial Coupling , 2012, ArXiv.

[71]  Stéphane Mallat,et al.  Solving Inverse Problems With Piecewise Linear Estimators: From Gaussian Mixture Models to Structured Sparsity , 2010, IEEE Transactions on Image Processing.

[72]  Arnaud Doucet,et al.  Asymptotic bias of stochastic gradient search , 2011, IEEE Conference on Decision and Control and European Control Conference.

[73]  B. Efron Tweedie’s Formula and Selection Bias , 2011, Journal of the American Statistical Association.

[74]  Yair Weiss,et al.  From learning models of natural image patches to whole image restoration , 2011, 2011 International Conference on Computer Vision.

[75]  Stephen P. Boyd,et al.  Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers , 2011, Found. Trends Mach. Learn..

[76]  Rémi Gribonval,et al.  Should Penalized Least Squares Regression be Interpreted as Maximum A Posteriori Estimation? , 2011, IEEE Transactions on Signal Processing.

[77]  Heinz H. Bauschke,et al.  Convex Analysis and Monotone Operator Theory in Hilbert Spaces , 2011, CMS Books in Mathematics.

[78]  M. Girolami,et al.  Riemann manifold Langevin and Hamiltonian Monte Carlo methods , 2011, Journal of the Royal Statistical Society: Series B (Statistical Methodology).

[79]  Yann LeCun,et al.  Learning Fast Approximations of Sparse Coding , 2010, ICML.

[80]  Andrew M. Stuart,et al.  Inverse problems: A Bayesian perspective , 2010, Acta Numerica.

[81]  Andrea Montanari,et al.  The dynamics of message passing on dense graphs, with applications to compressed sensing , 2010, 2010 IEEE International Symposium on Information Theory.

[82]  Andrea Montanari,et al.  Message-passing algorithms for compressed sensing , 2009, Proceedings of the National Academy of Sciences.

[83]  Alan C. Bovik,et al.  Mean squared error: Love it or leave it? A new look at Signal Fidelity Measures , 2009, IEEE Signal Processing Magazine.

[84]  Xiongzhi Chen Brownian Motion and Stochastic Calculus , 2008 .

[85]  Christian P. Robert,et al.  The Bayesian choice : from decision-theoretic foundations to computational implementation , 2007 .

[86]  Karen O. Egiazarian,et al.  Image denoising with block-matching and 3D filtering , 2006, Electronic Imaging.

[87]  Jean-Michel Morel,et al.  A non-local algorithm for image denoising , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[88]  Eero P. Simoncelli,et al.  Image quality assessment: from error visibility to structural similarity , 2004, IEEE Transactions on Image Processing.

[89]  M. Nikolova An Algorithm for Total Variation Minimization and Applications , 2004 .

[90]  É. Moulines,et al.  Convergence of a stochastic approximation version of the EM algorithm , 1999 .

[91]  R. Tweedie,et al.  Exponential convergence of Langevin distributions and their discrete approximations , 1996 .

[92]  B. Delyon General results on the convergence of stochastic algorithms , 1996, IEEE Trans. Autom. Control..

[93]  O. Brandière,et al.  Les algorithmes stochastiques contournent-ils les pièges? , 1995 .

[94]  Y. Marzouk,et al.  Large-Scale Inverse Problems and Quantification of Uncertainty , 1994 .

[95]  S. Meyn,et al.  Stability of Markovian processes III: Foster–Lyapunov criteria for continuous-time processes , 1993, Advances in Applied Probability.

[96]  L. Rudin,et al.  Nonlinear total variation based noise removal algorithms , 1992 .

[97]  M. Metivier,et al.  Applications of a Kushner and Clark lemma to general classes of stochastic algorithms , 1984, IEEE Trans. Inf. Theory.