Learning the optimal regularizer for inverse problems

In this work, we consider the linear inverse problem y = Ax+ ε, where A : X → Y is a known linear operator between the separable Hilbert spaces X and Y , x is a random variable inX and ε is a zero-mean random process in Y . This setting covers several inverse problems in imaging including denoising, deblurring and X-ray tomography. Within the classical framework of regularization, we focus on the case where the regularization functional is not given a priori, but learned from data. Our first result is a characterization of the optimal generalized Tikhonov regularizer, with respect to the mean squared error. We find that it is completely independent of the forward operator A and depends only on the mean and covariance of x. Then, we consider the problem of learning the regularizer from a finite training set in two different frameworks: one supervised, based on samples of both x and y, and one unsupervised, based only on samples of x. In both cases we prove generalization bounds, under some weak assumptions on the distribution of x and ε, including the case of sub-Gaussian variables. Our bounds hold in infinite-dimensional spaces, thereby showing that finer and finer discretizations do not make this learning problem harder. The results are validated through numerical simulations.

[1]  George Em Karniadakis,et al.  Error estimates for DeepOnets: A deep learning framework in infinite dimensions , 2021, Transactions of Mathematics and Its Applications.

[2]  Steven Kay,et al.  Fundamentals Of Statistical Signal Processing , 2001 .

[3]  Lea Fleischer,et al.  Regularization of Inverse Problems , 1996 .

[4]  O. Papaspiliopoulos High-Dimensional Probability: An Introduction with Applications in Data Science , 2020 .

[5]  Charles R. Johnson,et al.  Matrix analysis , 1985, Statistical Inference for Engineers and Data Scientists.

[6]  Carola-Bibiane Schönlieb,et al.  Bilevel approaches for learning of variational imaging models , 2015, ArXiv.

[7]  Matti Lassas,et al.  Analysis of regularized inversion of data corrupted by white Gaussian noise , 2013, 1311.6323.

[8]  Matti Lassas. Eero Saksman,et al.  Discretization-invariant Bayesian inversion and Besov space priors , 2009, 0901.4220.

[9]  Carola-Bibiane Schonlieb,et al.  Learned convex regularizers for inverse problems. , 2020 .

[10]  Simon R. Arridge,et al.  Solving inverse problems using data-driven models , 2019, Acta Numerica.

[11]  Jonas Adler,et al.  Learned Primal-Dual Reconstruction , 2017, IEEE Transactions on Medical Imaging.

[12]  Marco Prato,et al.  Deep neural networks for inverse problems with pseudodifferential operators: an application to limited-angle tomography , 2020, SIAM J. Imaging Sci..

[13]  Jonas Adler,et al.  Solving ill-posed inverse problems using iterative deep neural networks , 2017, ArXiv.

[14]  Jan MacDonald,et al.  Solving Inverse Problems With Deep Neural Networks - Robustness Included? , 2020, ArXiv.

[15]  Shai Ben-David,et al.  Understanding Machine Learning: From Theory to Algorithms , 2014 .

[16]  Rebecca Willett,et al.  Deep Equilibrium Architectures for Inverse Problems in Imaging , 2021, IEEE Transactions on Computational Imaging.

[17]  Felipe Cucker,et al.  On the mathematical foundations of learning , 2001 .

[18]  Maarten V. de Hoop,et al.  Learning Schatten-Von Neumann Operators , 2019, ArXiv.

[19]  Carola-Bibiane Schönlieb,et al.  Adversarially learned iterative reconstruction for imaging inverse problems , 2021, SSVM.

[20]  Alexandros G. Dimakis,et al.  Deep Learning Techniques for Inverse Problems in Imaging , 2020, IEEE Journal on Selected Areas in Information Theory.

[21]  O. Scherzer,et al.  Sparse regularization with lq penalty term , 2008, 0806.3222.

[22]  B. Carl,et al.  Entropy, Compactness and the Approximation of Operators , 1990 .

[23]  Carola-Bibiane Schönlieb,et al.  Adversarial Regularizers in Inverse Problems , 2018, NeurIPS.

[24]  Simon R. Arridge,et al.  Multi-Scale Learned Iterative Reconstruction , 2019, IEEE Transactions on Computational Imaging.

[25]  Andrew M. Stuart,et al.  Inverse problems: A Bayesian perspective , 2010, Acta Numerica.

[26]  V. Koltchinskii,et al.  Oracle inequalities in empirical risk minimization and sparse recovery problems , 2011 .

[27]  Thomas Pock,et al.  Learning a variational network for reconstruction of accelerated MRI data , 2017, Magnetic resonance in medicine.

[28]  Peter L. Bartlett,et al.  Rademacher and Gaussian Complexities: Risk Bounds and Structural Results , 2003, J. Mach. Learn. Res..

[29]  M. Salo,et al.  On instability mechanisms for inverse problems , 2020, 2012.01855.

[30]  M. D. Hoop,et al.  Deep learning architectures for nonlinear operator functions and nonlinear inverse problems , 2019, Mathematical Statistics and Learning.

[31]  V. Koltchinskii,et al.  Concentration inequalities and moment bounds for sample covariance operators , 2014, 1405.2468.

[32]  Faming Liang,et al.  Statistical and Computational Inverse Problems , 2006, Technometrics.

[33]  Joel Franklin,et al.  Well-posed stochastic extensions of ill-posed linear problems☆ , 1970 .

[34]  Karl Kunisch,et al.  Total Deep Variation for Linear Inverse Problems , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[35]  Thomas Pock,et al.  Variational Networks: Connecting Variational Methods and Deep Learning , 2017, GCPR.