MMSE Bounds Under Kullback–Leibler Divergence Constraints on the Joint Input-Output Distribution

This paper proposes a new family of lower and upper bounds on the minimum mean squared error (MMSE). The key idea is to minimize/maximize the MMSE subject to the constraint that the joint distribution of the input-output statistics lies in a Kullback-Leibler divergence ball centered at some Gaussian reference distribution. Both bounds are tight and are attained by Gaussian distributions whose mean is identical to that of the reference distribution and whose covariance matrix is determined by a scalar parameter that can be obtained by finding the root of a monotonic function. The upper bound corresponds to a minimax optimal estimator and provides performance guarantees under distributional uncertainty. The lower bound provides an alternative to well-known inequalities in estimation theory, such as the Cramer-Rao bound, that is potentially tighter and defined for a larger class of distributions. Examples of applications in signal processing and information theory illustrate the usefulness of the proposed bounds in practice.

[1]  Shlomo Shamai,et al.  Mutual information and minimum mean-square error in Gaussian channels , 2004, IEEE Transactions on Information Theory.

[2]  Edward R. Dougherty,et al.  Exact Sample Conditioned MSE Performance of the Bayesian MMSE Estimator for Classification Error—Part I: Representation , 2012, IEEE Transactions on Signal Processing.

[3]  Edward R. Dougherty,et al.  Exact Sample Conditioned MSE Performance of the Bayesian MMSE Estimator for Classification Error—Part II: Consistency and Performance Analysis , 2012, IEEE Transactions on Signal Processing.

[4]  M. Raginsky Concentration of Measure Inequalities in Information Theory, Communications, and Coding: Second Edition , 2014 .

[5]  S. Li Concise Formulas for the Area and Volume of a Hyperspherical Cap , 2011 .

[6]  Michael Muma,et al.  Robust Statistics for Signal Processing , 2018 .

[7]  H. Vincent Poor,et al.  A Class of Lower Bounds for Bayesian Risk with a Bregman Loss , 2020, 2020 IEEE 21st International Workshop on Signal Processing Advances in Wireless Communications (SPAWC).

[8]  Sergio Verdú,et al.  A simple proof of the entropy-power inequality , 2006, IEEE Transactions on Information Theory.

[9]  Shlomo Shamai,et al.  Estimation in Gaussian Noise: Properties of the Minimum Mean-Square Error , 2010, IEEE Transactions on Information Theory.

[10]  Michael Muma,et al.  Robust Estimation in Signal Processing: A Tutorial-Style Treatment of Fundamental Concepts , 2012, IEEE Signal Processing Magazine.

[11]  S. Kay Fundamentals of statistical signal processing: estimation theory , 1993 .

[12]  H. Poor,et al.  Robust system‐parameter identification: The influence functional approach , 1992 .

[13]  H. Vincent Poor,et al.  On Communication Through a Gaussian Channel With an MMSE Disturbance Constraint , 2016, IEEE Transactions on Information Theory.

[14]  Jacob Ziv,et al.  Some lower bounds on signal parameter estimation , 1969, IEEE Trans. Inf. Theory.

[15]  J. Goodier The Concise Encyclopedia of Statistics , 2009 .

[16]  Shlomo Shamai,et al.  On the Structure of the Least Favorable Prior Distributions , 2018, 2018 IEEE International Symposium on Information Theory (ISIT).

[17]  Abdelhak M. Zoubir,et al.  MMSE Bounds for Additive Noise Channels Under Kullback–Leibler Divergence Constraints on the Input Distribution , 2019, IEEE Transactions on Signal Processing.

[18]  Ehud Weinstein,et al.  A general class of lower bounds in parameter estimation , 1988, IEEE Trans. Inf. Theory.

[19]  Stefan Aachen,et al.  Signal Detection In Non Gaussian Noise , 2016 .

[20]  Venkat Anantharam,et al.  A variational characterization of Rényi divergences , 2017, 2017 IEEE International Symposium on Information Theory (ISIT).

[21]  Kimmo Kansanen,et al.  On MMSE Estimation: A Linear Model Under Gaussian Mixture Statistics , 2012, IEEE Transactions on Signal Processing.

[22]  Shlomo Shamai,et al.  On additive channels with generalized Gaussian noise , 2017, 2017 IEEE International Symposium on Information Theory (ISIT).

[23]  Avraham Adler,et al.  Lambert-W Function , 2015 .

[24]  Igal Sason,et al.  Concentration of Measure Inequalities in Information Theory, Communications, and Coding , 2012, Found. Trends Commun. Inf. Theory.

[25]  P. Davis Leonhard Euler's Integral: A Historical Profile of the Gamma Function: In Memoriam: Milton Abramowitz , 1959 .

[26]  C. Jennison,et al.  Robust Statistics: The Approach Based on Influence Functions , 1987 .