A Vector Generalization of Costa's Entropy-Power Inequality With Applications

This paper considers an entropy-power inequality (EPI) of Costa and presents a natural vector generalization with a real positive semidefinite matrix parameter. The new inequality is proved using a perturbation approach via a fundamental relationship between the derivative of mutual information and the minimum mean-square error (MMSE) estimate in linear vector Gaussian channels. As an application, a new extremal entropy inequality is derived from the generalized Costa EPI and then used to establish the secrecy capacity regions of the degraded vector Gaussian broadcast channel with layered confidential messages.

[1]  L. Ozarow,et al.  On a source-coding problem with two channels and three receivers , 1980, The Bell System Technical Journal.

[2]  Shlomo Shamai,et al.  Mutual information and minimum mean-square error in Gaussian channels , 2004, IEEE Transactions on Information Theory.

[3]  Amir K. Khandani,et al.  Secrecy Rate Region of the Broadcast Channel with an Eavesdropper , 2009, ArXiv.

[4]  Tie Liu,et al.  An Extremal Inequality Motivated by Multiterminal Information-Theoretic Problems , 2006, IEEE Transactions on Information Theory.

[5]  Max H. M. Costa,et al.  A new entropy power inequality , 1985, IEEE Trans. Inf. Theory.

[6]  A. Banerjee Convex Analysis and Optimization , 2006 .

[7]  George A. F. Seber,et al.  A matrix handbook for statisticians , 2007 .

[8]  Katalin Marton,et al.  A coding theorem for the discrete memoryless broadcast channel , 1979, IEEE Trans. Inf. Theory.

[9]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[10]  Amir Dembo,et al.  Information theoretic inequalities , 1991, IEEE Trans. Inf. Theory.

[11]  Shlomo Shamai,et al.  Proof of Entropy Power Inequalities Via MMSE , 2006, 2006 IEEE International Symposium on Information Theory.

[12]  Audra E. Kosh,et al.  Linear Algebra and its Applications , 1992 .

[13]  Patrick P. Bergmans,et al.  Random coding theorem for broadcast channels with degraded components , 1973, IEEE Trans. Inf. Theory.

[14]  Imre Csiszár,et al.  Broadcast channels with confidential messages , 1978, IEEE Trans. Inf. Theory.

[15]  Sang Joon Kim,et al.  A Mathematical Theory of Communication , 2006 .

[16]  Shlomo Shamai,et al.  A vector generalization of Costa entropy-power inequality and applications , 2009, 2009 IEEE International Symposium on Information Theory.

[17]  Amos Lapidoth,et al.  Capacity bounds via duality with applications to multiple-antenna systems on flat-fading channels , 2003, IEEE Trans. Inf. Theory.

[18]  Yasutada Oohama,et al.  The Rate-Distortion Function for the Quadratic Gaussian CEO Problem , 1998, IEEE Trans. Inf. Theory.

[19]  A.K. Khandani,et al.  Secure broadcasting : The secrecy rate region , 2008, 2008 46th Annual Allerton Conference on Communication, Control, and Computing.

[20]  A. D. Wyner,et al.  The wire-tap channel , 1975, The Bell System Technical Journal.

[21]  Daniel Pérez Palomar,et al.  Gradient of mutual information in linear vector Gaussian channels , 2006, IEEE Transactions on Information Theory.

[22]  Sennur Ulukus,et al.  The Secrecy Capacity Region of the Gaussian MIMO Multi-Receiver Wiretap Channel , 2009, IEEE Transactions on Information Theory.

[23]  Shlomo Shamai,et al.  A Note on the Secrecy Capacity of the Multiple-Antenna Wiretap Channel , 2007, IEEE Transactions on Information Theory.

[24]  Daniel Pérez Palomar,et al.  Gradient of mutual information in linear vector Gaussian channels , 2005, ISIT.

[25]  Sriram Vishwanath,et al.  On the Capacity of One-Sided Two User Gaussian Fading Broadcast Channels , 2008, IEEE GLOBECOM 2008 - 2008 IEEE Global Telecommunications Conference.

[26]  Jun Chen,et al.  Rate Region of Gaussian Multiple Description Coding With Individual and Central Distortion Constraints , 2009, IEEE Transactions on Information Theory.

[27]  A. J. Stam Some Inequalities Satisfied by the Quantities of Information of Fisher and Shannon , 1959, Inf. Control..

[28]  Amir Dembo,et al.  Simple proof of the concavity of the entropy power with respect to Gaussian noise , 1989, IEEE Trans. Inf. Theory.

[29]  Suhas N. Diggavi,et al.  The worst additive noise under a covariance constraint , 2001, IEEE Trans. Inf. Theory.

[30]  Max H. M. Costa,et al.  On the Gaussian interference channel , 1985, IEEE Trans. Inf. Theory.

[31]  Daniel Pérez Palomar,et al.  Hessian and Concavity of Mutual Information, Differential Entropy, and Entropy Power in Linear Vector Gaussian Channels , 2009, IEEE Transactions on Information Theory.

[32]  Martin E. Hellman,et al.  The Gaussian wire-tap channel , 1978, IEEE Trans. Inf. Theory.

[33]  Shlomo Shamai,et al.  The Capacity Region of the Gaussian Multiple-Input Multiple-Output Broadcast Channel , 2006, IEEE Transactions on Information Theory.