Monotonicity, Thinning, and Discrete Versions of the Entropy Power Inequality

We consider the entropy of sums of independent discrete random variables, in analogy with Shannon's Entropy Power Inequality, where equality holds for normals. In our case, infinite divisibility suggests that equality should hold for Poisson variables. We show that some natural analogues of the EPI do not in fact hold, but propose an alternative formulation which does always hold. The key to many proofs of Shannon's EPI is the behavior of entropy on scaling of continuous random variables. We believe that Rényi's operation of thinning discrete random variables plays a similar role to scaling, and give a sharp bound on how the entropy of ultra log-concave random variables behaves on thinning. In the spirit of the monotonicity results established by Artstein, Ball, Barthe, and Naor, we prove a stronger version of concavity of entropy, which implies a strengthened form of our discrete EPI.

[1]  Kevin Barraclough,et al.  I and i , 2001, BMJ : British Medical Journal.

[2]  Antonia Maria Tulino,et al.  Monotonic Decrease of the Non-Gaussianness of the Sum of Independent Random Variables: A Simple Proof , 2006, IEEE Transactions on Information Theory.

[3]  M. B. Porter On the Differentiation of An Infinite Series Term by Term , 1901 .

[4]  Sang Joon Kim,et al.  A Mathematical Theory of Communication , 2006 .

[5]  W. Beckner Inequalities in Fourier analysis , 1975 .

[6]  Amir Dembo,et al.  Information theoretic inequalities , 1991, IEEE Trans. Inf. Theory.

[7]  Nelson M. Blachman,et al.  The convolution inequality for entropy powers , 1965, IEEE Trans. Inf. Theory.

[8]  Mokshay M. Madiman,et al.  Generalized Entropy Power Inequalities and Monotonicity Properties of Information , 2006, IEEE Transactions on Information Theory.

[9]  Oliver Johnson,et al.  Thinning, Entropy, and the Law of Thin Numbers , 2009, IEEE Transactions on Information Theory.

[10]  Claude E. Shannon,et al.  A mathematical theory of communication , 1948, MOCO.

[11]  W. Marsden I and J , 2012 .

[12]  E. Lieb Proof of an entropy conjecture of Wehrl , 1978 .

[13]  Shlomo Shamai,et al.  A binary analog to the entropy-power inequality , 1990, IEEE Trans. Inf. Theory.

[14]  Hans S. Witsenhausen,et al.  Entropy inequalities for discrete channels , 1974, IEEE Trans. Inf. Theory.

[15]  O. Johnson Log-concavity and the maximum entropy property of the Poisson distribution , 2006, math/0603647.

[16]  F. Baccelli,et al.  Characterization of Poisson Processes , 1987 .

[17]  K. Ball,et al.  Solution of Shannon's problem on the monotonicity of entropy , 2004 .

[18]  Aaron D. Wyner,et al.  A theorem on the entropy of certain binary sequences and applications-II , 1973, IEEE Trans. Inf. Theory.

[19]  J. Kingman Uses of Exchangeability , 1978 .

[20]  A. J. Stam Some Inequalities Satisfied by the Quantities of Information of Fisher and Shannon , 1959, Inf. Control..

[21]  R. Pemantle Towards a theory of negative dependence , 2000, math/0404095.

[22]  B. Gnedenko,et al.  Random Summation: Limit Theorems and Applications , 1996 .

[23]  Thomas M. Liggett Ultra Logconcave Sequences and Negative Dependence , 1997, J. Comb. Theory, Ser. A.

[24]  Oliver Johnson,et al.  Concavity of entropy under thinning , 2009, 2009 IEEE International Symposium on Information Theory.

[25]  Aaron D. Wyner,et al.  A theorem on the entropy of certain binary sequences and applications-I , 1973, IEEE Trans. Inf. Theory.

[26]  Liming Wu,et al.  A new modified logarithmic Sobolev inequality for Poisson point processes and several applications , 2000 .

[27]  Oliver Johnson,et al.  Thinning and the Law of Small Numbers , 2007, 2007 IEEE International Symposium on Information Theory.

[28]  Ingram Olkin,et al.  Entropy of the Sum of Independent Bernoulli Random Variables and of the Multinomial Distribution , 1981 .

[29]  Sergio Verdú,et al.  A simple proof of the entropy-power inequality , 2006, IEEE Transactions on Information Theory.

[30]  Yaming Yu,et al.  Monotonic Convergence in an Information-Theoretic Law of Small Numbers , 2008, IEEE Transactions on Information Theory.

[31]  Max H. M. Costa,et al.  A new entropy power inequality , 1985, IEEE Trans. Inf. Theory.

[32]  Christophe Vignat,et al.  AN ENTROPY POWER INEQUALITY FOR THE BINOMIAL FAMILY , 2003 .