Sampling Theorems for Learning from Incomplete Measurements

In many real-world settings, only incomplete measurement data are available which can pose a problem for learning. Unsupervised learning of the signal model using a fixed incomplete measurement process is impossible in general, as there is no information in the nullspace of the measurement operator. This limitation can be overcome by using measurements from multiple operators. While this idea has been successfully applied in various applications, a precise characterization of the conditions for learning is still lacking. In this paper, we fill this gap by presenting necessary and sufficient conditions for learning the signal model which indicate the interplay between the number of distinct measurement operators G, the number of measurements per operator m, the dimension of the model k and the dimension of the signals n. In particular, we show that generically unsupervised learning is possible if each operator obtains at least m > k + n/G measurements. Our results are agnostic of the learning algorithm and have implications in a wide range of practical algorithms, from low-rank matrix recovery to deep neural networks.

[1]  Mike E. Davies,et al.  Equivariant Imaging: Learning Beyond the Range Space , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[2]  Vivek K Goyal,et al.  Advances in Single-Photon Lidar for Autonomous Vehicles: Working Principles, Challenges, and Recent Advances , 2020, IEEE Signal Processing Magazine.

[3]  Ulugbek S. Kamilov,et al.  RARE: Image Reconstruction Using Deep Priors Learned Without Groundtruth , 2019, IEEE Journal of Selected Topics in Signal Processing.

[4]  Gordon Wetzstein,et al.  Confocal non-line-of-sight imaging based on the light-cone transform , 2018, Nature.

[5]  Alexandros G. Dimakis,et al.  AmbientGAN: Generative models from lossy measurements , 2018, ICLR.

[6]  Andrea Vedaldi,et al.  Deep Image Prior , 2017, International Journal of Computer Vision.

[7]  Michael Unser,et al.  Deep Convolutional Neural Network for Inverse Problems in Imaging , 2016, IEEE Transactions on Image Processing.

[8]  Robert D. Nowak,et al.  The Information-Theoretic Requirements of Subspace Clustering with Missing Data , 2016, ICML.

[9]  Hayder Radha,et al.  New guarantees for Blind Compressed Sensing , 2015, 2015 53rd Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[10]  Daniel L. Pimentel-Alarcón,et al.  A characterization of deterministic sampling patterns for low-rank matrix completion , 2015, 2015 53rd Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[11]  Guillermo Sapiro,et al.  Compressive Sensing by Learning a Gaussian Mixture Model From Measurements , 2015, IEEE Transactions on Image Processing.

[12]  Patrick Pérez,et al.  Fundamental Performance Limits for Ideal Decoders in High-Dimensional Linear Inverse Problems , 2013, IEEE Transactions on Information Theory.

[13]  Andrea J. Goldsmith,et al.  Exact and Stable Covariance Estimation From Quadratic Sampling via Convex Programming , 2013, IEEE Transactions on Information Theory.

[14]  Anders P. Eriksson,et al.  Fast Convolutional Sparse Coding , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition.

[15]  Richard G. Baraniuk,et al.  Dictionary learning from sparsely corrupted or compressed signals , 2012, 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[16]  Robert D. Nowak,et al.  High-Rank Matrix Completion , 2012, AISTATS.

[17]  Yonina C. Eldar,et al.  Blind Compressed Sensing Over a Structured Union of Subspaces , 2011, ArXiv.

[18]  Yonina C. Eldar,et al.  Blind Compressed Sensing , 2010, IEEE Transactions on Information Theory.

[19]  Mike E. Davies,et al.  Sampling Theorems for Signals From the Union of Finite-Dimensional Linear Subspaces , 2009, IEEE Transactions on Information Theory.

[20]  Emmanuel J. Candès,et al.  A Singular Value Thresholding Algorithm for Matrix Completion , 2008, SIAM J. Optim..

[21]  Emmanuel J. Candès,et al.  Exact Matrix Completion via Convex Optimization , 2008, Found. Comput. Math..

[22]  Matthias Hein,et al.  Intrinsic dimensionality estimation of submanifolds in Rd , 2005, ICML.

[23]  L. Rudin,et al.  Nonlinear total variation based noise removal algorithms , 1992 .

[24]  H. Wold,et al.  Some Theorems on Distribution Functions , 1936 .

[25]  Junqi Tang,et al.  Supplementary Materials for The Neural Tangent Link Between CNN Denoisers and Non-Local Filters , 2021 .

[26]  Michael Unser,et al.  Multi-CryoGAN: Reconstruction of Continuous Conformations in Cryo-EM Using Generative Adversarial Networks , 2020, ECCV Workshops.