Data-Processing Inequalities Based on a Certain Structured Class of Information Measures With Application to Estimation Theory

We study data-processing inequalities that are derived from a certain class of generalized information measures, where a series of convex functions and multiplicative likelihood ratios is nested alternately. While these information measures can be viewed as a special case of the most general Zakai-Ziv generalized information measure, this special nested structure calls for attention and motivates our study. Specifically, a certain choice of the convex functions leads to an information measure that extends the notion of the Bhattacharyya distance (or the Chernoff divergence): While the ordinary Bhattacharyya distance is based on the (weighted) geometric mean of two replicas of the channel's conditional distribution, the more general information measure allows an arbitrary number of such replicas. We apply the data-processing inequality induced by this information measure to a detailed study of lower bounds of parameter estimation under additive white Gaussian noise (AWGN) and show that in certain cases, tighter bounds can be obtained by using more than two replicas. While the resulting lower bound may not compete favorably with the best bounds available for the ordinary AWGN channel, the advantage of the new lower bound, relative to the other bounds, becomes significant in the presence of channel uncertainty, like unknown fading. This different behavior in the presence of channel uncertainty is explained by the convexity property of the information measure.

[1]  Andrew J. Viterbi,et al.  Principles of Digital Communication and Coding , 1979 .

[2]  Jacob Ziv,et al.  Some lower bounds on signal parameter estimation , 1969, IEEE Trans. Inf. Theory.

[3]  Neri Merhav Threshold Effects in Parameter Estimation as Phase Transitions in Statistical Mechanics , 2011, IEEE Trans. Inf. Theory.

[4]  Suguru Arimoto,et al.  On the converse to the coding theorem for discrete memoryless channels (Corresp.) , 1973, IEEE Trans. Inf. Theory.

[5]  Ehud Weinstein,et al.  A lower bound on the mean-square error in random parameter estimation , 1985, IEEE Trans. Inf. Theory.

[6]  R. Gallager Information Theory and Reliable Communication , 1968 .

[7]  S. Shamai,et al.  Information rates and error exponents of compound channels with application to antipodal signaling in a fading environment , 1993 .

[8]  Jacob Ziv,et al.  Improved Lower Bounds on Signal Parameter Estimation , 1975, IEEE Trans. Inf. Theory.

[9]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[10]  E. Weinstein,et al.  Lower bounds on the mean square estimation error , 1985, Proceedings of the IEEE.

[11]  C. Tellambura,et al.  Derivation of Craig's formula for Gaussian probability function , 1999 .

[12]  J. Ziv,et al.  A Generalization of the Rate-Distortion Theory and Applications , 1975 .

[13]  Neri Merhav,et al.  Data Processing Theorems and the Second Law of Thermodynamics , 2010, IEEE Transactions on Information Theory.

[14]  I. M. Jacobs,et al.  Principles of Communication Engineering , 1965 .

[15]  Neri Merhav,et al.  On the threshold effect in the estimation of chaotic sequences , 2003, IEEE Transactions on Information Theory.

[16]  Jacob Ziv,et al.  On functionals satisfying a data-processing theorem , 1973, IEEE Trans. Inf. Theory.

[17]  Imre Csiszár,et al.  Information Theory and Statistics: A Tutorial , 2004, Found. Trends Commun. Inf. Theory.

[18]  I. Csiszár A class of measures of informativity of observation channels , 1972 .

[19]  Stephen P. Boyd,et al.  Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.

[20]  Rodger E. Ziemer,et al.  Principles of Communication , 2001 .