Classification and Reconstruction of High-Dimensional Signals From Low-Dimensional Features in the Presence of Side Information

This paper offers a characterization of fundamental limits on the classification and reconstruction of high-dimensional signals from low-dimensional features, in the presence of side information. We consider a scenario where a decoder has access both to linear features of the signal of interest and to linear features of the side information signal; while the side information may be in a compressed form, the objective is recovery or classification of the primary signal, not the side information. The signal of interest and the side information are each assumed to have (distinct) latent discrete labels; conditioned on these two labels, the signal of interest and side information are drawn from a multivariate Gaussian distribution that correlates the two. With joint probabilities on the latent labels, the overall signal-(side information) representation is defined by a Gaussian mixture model. By considering bounds to the misclassification probability associated with the recovery of the underlying signal label, and bounds to the reconstruction error associated with the recovery of the signal of interest itself, we then provide sharp sufficient and/or necessary conditions for these quantities to approach zero when the covariance matrices of the Gaussians are nearly low rank. These conditions, which are reminiscent of the well-known Slepian-Wolf and Wyner-Ziv conditions, are the function of the number of linear features extracted from signal of interest, the number of linear features extracted from the side information signal, and the geometry of these signals and their interplay. Moreover, on assuming that the signal of interest and the side information obey such an approximately low-rank model, we derive the expansions of the reconstruction error as a function of the deviation from an exactly low-rank model; such expansions also allow the identification of operational regimes, where the impact of side information on signal reconstruction is most relevant. Our framework, which offers a principled mechanism to integrate side information in high-dimensional data problems, is also tested in the context of imaging applications. In particular, we report state-of-theart results in compressive hyperspectral imaging applications, where the accompanying side information is a conventional digital photograph.

[1]  H. Sorenson,et al.  Recursive bayesian estimation using gaussian sums , 1971 .

[2]  Jack K. Wolf,et al.  Noiseless coding of correlated information sources , 1973, IEEE Trans. Inf. Theory.

[3]  Adi Ben-Israel,et al.  Generalized inverses: theory and applications , 1974 .

[4]  Rudolf Ahlswede,et al.  Source coding with side information and a converse for degraded broadcast channels , 1975, IEEE Trans. Inf. Theory.

[5]  Aaron D. Wyner,et al.  The rate-distortion function for source coding with side information at the decoder , 1976, IEEE Trans. Inf. Theory.

[6]  Diane Valérie Ouellette Schur complements and statistics , 1981 .

[7]  M. Saunders,et al.  Towards a Generalized Singular Value Decomposition , 1981 .

[8]  Charles R. Johnson,et al.  Matrix analysis , 1985, Statistical Inference for Engineers and Data Scientists.

[9]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[10]  Stéphane Mallat,et al.  Matching pursuits with time-frequency dictionaries , 1993, IEEE Trans. Signal Process..

[11]  A. Robert Calderbank,et al.  Space-Time Codes for High Data Rate Wireless Communications : Performance criterion and Code Construction , 1998, IEEE Trans. Inf. Theory.

[12]  Michael A. Saunders,et al.  Atomic Decomposition by Basis Pursuit , 1998, SIAM J. Sci. Comput..

[13]  Robert D. Nowak,et al.  Wavelet-based statistical signal processing using hidden Markov models , 1998, IEEE Trans. Signal Process..

[14]  A. Robert Calderbank,et al.  Space-Time block codes from orthogonal designs , 1999, IEEE Trans. Inf. Theory.

[15]  Fuzhen Zhang,et al.  Some inequalities on generalized Schur complements , 1999 .

[16]  David G. Stork,et al.  Pattern Classification (2nd ed.) , 1999 .

[17]  J. Tenenbaum,et al.  A global geometric framework for nonlinear dimensionality reduction. , 2000, Science.

[18]  Anil K. Jain,et al.  Statistical Pattern Recognition: A Review , 2000, IEEE Trans. Pattern Anal. Mach. Intell..

[19]  Carl D. Meyer,et al.  Matrix Analysis and Applied Linear Algebra , 2000 .

[20]  David G. Stork,et al.  Pattern classification, 2nd Edition , 2000 .

[21]  Shigeo Abe DrEng Pattern Classification , 2001, Springer London.

[22]  Kari Torkkola,et al.  Learning Discriminative Feature Transforms to Low Dimensions in Low Dimentions , 2001, NIPS.

[23]  Petra Perner,et al.  Data Mining - Concepts and Techniques , 2002, Künstliche Intell..

[24]  Kari Torkkola,et al.  Feature Extraction by Non-Parametric Mutual Information Maximization , 2003, J. Mach. Learn. Res..

[25]  Isabelle Guyon,et al.  An Introduction to Variable and Feature Selection , 2003, J. Mach. Learn. Res..

[26]  Samuel Kaski,et al.  Informative Discriminant Analysis , 2003, ICML.

[27]  Deniz Erdogmus,et al.  Lower and Upper Bounds for Misclassification Probability Based on Renyi's Information , 2004, J. VLSI Signal Process..

[28]  E. Candès,et al.  Stable signal recovery from incomplete and inaccurate measurements , 2005, math/0503066.

[29]  Emmanuel J. Candès,et al.  Decoding by linear programming , 2005, IEEE Transactions on Information Theory.

[30]  G. Styan,et al.  Rank equalities for idempotent matrices with applications , 2006 .

[31]  D. Foster,et al.  Frequency of metamerism in natural scenes. , 2007, Journal of the Optical Society of America. A, Optics, image science, and vision.

[32]  Deniz Erdogmus,et al.  Feature extraction using information-theoretic learning , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[33]  Emmanuel J. Candès,et al.  Robust uncertainty principles: exact signal reconstruction from highly incomplete frequency information , 2004, IEEE Transactions on Information Theory.

[34]  Pietro Perona,et al.  One-shot learning of object categories , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[35]  Christopher M. Bishop,et al.  Pattern Recognition and Machine Learning (Information Science and Statistics) , 2006 .

[36]  Kaare Brandt Petersen,et al.  The Matrix Cookbook , 2006 .

[37]  Emmanuel J. Candès,et al.  Near-Optimal Signal Recovery From Random Projections: Universal Encoding Strategies? , 2004, IEEE Transactions on Information Theory.

[38]  Zoran Nenadic,et al.  Information Discriminant Analysis: Feature Extraction with an Information-Theoretic Objective , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[39]  Jie Tang,et al.  Prior image constrained compressed sensing (PICCS): a method to accurately reconstruct dynamic CT images from highly undersampled projection data sets. , 2008, Medical physics.

[40]  Lawrence Carin,et al.  Bayesian Compressive Sensing , 2008, IEEE Transactions on Signal Processing.

[41]  R. DeVore,et al.  A Simple Proof of the Restricted Isometry Property for Random Matrices , 2008 .

[42]  Richard G. Baraniuk,et al.  Random Projections of Smooth Manifolds , 2009, Found. Comput. Math..

[43]  Xuelong Li,et al.  Geometric Mean for Subspace Selection , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[44]  Babak Hassibi,et al.  On the Reconstruction of Block-Sparse Signals With an Optimal Number of Measurements , 2008, IEEE Transactions on Signal Processing.

[45]  Yonina C. Eldar,et al.  Robust Recovery of Signals From a Structured Union of Subspaces , 2008, IEEE Transactions on Information Theory.

[46]  Richard G. Baraniuk,et al.  Distributed Compressive Sensing , 2009, ArXiv.

[47]  David B. Dunson,et al.  Multitask Compressive Sensing , 2009, IEEE Transactions on Signal Processing.

[48]  Mike E. Davies,et al.  Sampling Theorems for Signals From the Union of Finite-Dimensional Linear Subspaces , 2009, IEEE Transactions on Information Theory.

[49]  Allen Y. Yang,et al.  Robust Face Recognition via Sparse Representation , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[50]  David J. Brady,et al.  Multiframe image estimation for coded aperture snapshot spectral imagers. , 2010, Applied optics.

[51]  Wei Lu,et al.  Modified-CS: Modifying compressive sensing for problems with partially known support , 2009, 2009 IEEE International Symposium on Information Theory.

[52]  Richard G. Baraniuk,et al.  Signal Processing With Compressive Measurements , 2010, IEEE Journal of Selected Topics in Signal Processing.

[53]  Stephen J. Wright,et al.  Computational Methods for Sparse Solution of Linear Inverse Problems , 2010, Proceedings of the IEEE.

[54]  Yonina C. Eldar,et al.  Block-Sparse Signals: Uncertainty Relations and Efficient Recovery , 2009, IEEE Transactions on Signal Processing.

[55]  Volkan Cevher,et al.  Model-Based Compressive Sensing , 2008, IEEE Transactions on Information Theory.

[56]  Martin Vetterli,et al.  Distributed Sampling of Signals Linked by Sparse Filtering: Theory and Applications , 2010, IEEE Transactions on Signal Processing.

[57]  David B. Dunson,et al.  Compressive Sensing on Manifolds Using a Nonparametric Mixture of Factor Analyzers: Algorithm and Performance Bounds , 2010, IEEE Transactions on Signal Processing.

[58]  Allen Y. Yang,et al.  Distributed Sensor Perception via Sparse Representation , 2010, Proceedings of the IEEE.

[59]  Guillermo Sapiro,et al.  Statistical Compressed Sensing of Gaussian Mixture Models , 2011, IEEE Transactions on Signal Processing.

[60]  A. Robert Calderbank,et al.  Communications Inspired Linear Discriminant Analysis , 2012, ICML.

[61]  Michael B. Wakin,et al.  The Restricted Isometry Property for Random Block Diagonal Matrices , 2012, ArXiv.

[62]  Paul W. Fieguth,et al.  Texture Classification from Random Features , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[63]  Stéphane Mallat,et al.  Solving Inverse Problems With Piecewise Linear Estimators: From Gaussian Mixture Models to Structured Sparsity , 2010, IEEE Transactions on Image Processing.

[64]  A. Robert Calderbank,et al.  Communications-Inspired Projection Design with Application to Compressive Sensing , 2012, SIAM J. Imaging Sci..

[65]  A. Robert Calderbank,et al.  Projections designs for compressive classification , 2013, 2013 IEEE Global Conference on Signal and Information Processing.

[66]  Richard G. Baraniuk,et al.  Measurement Bounds for Sparse Signal Ensembles via Graphical Models , 2011, IEEE Transactions on Information Theory.

[67]  A. Robert Calderbank,et al.  Compressive classification , 2013, 2013 IEEE International Symposium on Information Theory.

[68]  R. Gribonval,et al.  Exact Recovery Conditions for Sparse Representations With Partial Support Information , 2013, IEEE Transactions on Information Theory.

[69]  Lawrence Carin,et al.  Coded Hyperspectral Imaging and Blind Compressive Sensing , 2013, SIAM J. Imaging Sci..

[70]  A. Robert Calderbank,et al.  Designed Measurements for Vector Count Data , 2013, NIPS.

[71]  Guillermo Sapiro,et al.  Ieee Transactions on Signal Processing Task-driven Adaptive Statistical Compressive Sensing of Gaussian Mixture Models Ieee Transactions on Signal Processing 2 , 2022 .

[72]  Adel Javanmard,et al.  Information-Theoretically Optimal Compressed Sensing via Spatial Coupling and Approximate Message Passing , 2011, IEEE Transactions on Information Theory.

[73]  Ben Adcock,et al.  On asymptotic structure in compressed sensing , 2014, ArXiv.

[74]  A. Robert Calderbank,et al.  Reconstruction of Signals Drawn From a Gaussian Mixture Via Noisy Compressive Measurements , 2013, IEEE Transactions on Signal Processing.

[75]  Yonina C. Eldar,et al.  The application of Compressed Sensing for Longitudinal MRI , 2014, ArXiv.

[76]  Saeid Haghighatshoar Multi terminal probabilistic compressed sensing , 2014, 2014 IEEE International Symposium on Information Theory.

[77]  A. Robert Calderbank,et al.  A Bregman Matrix and the Gradient of Mutual Information for Vector Poisson and Gaussian Channels , 2014, IEEE Transactions on Information Theory.

[78]  Miguel R. D. Rodrigues,et al.  Compressed Sensing with Prior Information: Optimal Strategies, Geometry, and Bounds , 2014, ArXiv.

[79]  Guillermo Sapiro,et al.  Video Compressive Sensing Using Gaussian Mixture Models , 2014, IEEE Transactions on Image Processing.

[80]  Xing Wang,et al.  Side information-aided compressed sensing reconstruction via approximate message passing , 2013, 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[81]  A. Robert Calderbank,et al.  Nonlinear Information-Theoretic Compressive Measurement Design , 2014, ICML.

[82]  A. Robert Calderbank,et al.  Compressive Classification of a Mixture of Gaussians: Analysis, Designs and Geometrical Interpretation , 2014, ArXiv.

[83]  Volkan Cevher,et al.  Dynamic sparse state estimation using ℓ1-ℓ1 minimization: Adaptive-rate measurement bounds, algorithms and applications , 2015, 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[84]  Yonina C. Eldar,et al.  Compressed sensing for longitudinal MRI: An adaptive-weighted approach. , 2014, Medical physics.

[85]  Volkan Cevher,et al.  Adaptive-Rate Sparse Signal Reconstruction With Application in Compressive Background Subtraction , 2015, ArXiv.

[86]  Xing Wang,et al.  Approximate message passing-based compressed sensing reconstruction with generalized elastic net prior , 2013, Signal Process. Image Commun..

[87]  S. Frick,et al.  Compressed Sensing , 2014, Computer Vision, A Reference Guide.

[88]  Enrico Magli,et al.  Distributed Compressed Sensing , 2015 .

[89]  Ben Adcock,et al.  BREAKING THE COHERENCE BARRIER: A NEW THEORY FOR COMPRESSED SENSING , 2013, Forum of Mathematics, Sigma.