Bounds on variance for symmetric unimodal distributions

We show a direct relationship between the variance and the differential entropy for the general class of symmetric unimodal distributions by providing an upper bound on variance in terms of entropy power. Combining this bound with the well-known entropy power lower bound on variance, we prove that for the general class of symmetric unimodal distributions the variance can be bounded below and above by the scaled entropy power. As differential entropy decreases, the variance is sandwiched between two exponentially decreasing functions in the differential entropy. This establishes that for the general class of symmetric unimodal distributions, the differential entropy can be used as a surrogate for concentration of the distribution.

[1]  P. Groeneboom Estimating a monotone density , 1984 .

[2]  L. Birge,et al.  Estimation of unimodal densities without smoothness assumptions , 1997 .

[3]  Feller William,et al.  An Introduction To Probability Theory And Its Applications , 1950 .

[4]  Donald Geman,et al.  An Active Testing Model for Tracking Roads in Satellite Images , 1996, IEEE Trans. Pattern Anal. Mach. Intell..

[5]  Peter I. Frazier,et al.  Twenty Questions with Noise: Bayes Optimal Policies for Entropy Loss , 2012, Journal of Applied Probability.

[6]  M. Bagnoli,et al.  Log-concave probability and its applications , 2004 .

[7]  J. Hartigan,et al.  The Dip Test of Unimodality , 1985 .

[8]  G. Walther Inference and Modeling with Log-concave Distributions , 2009, 1010.0305.

[9]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[10]  S. Bobkov Isoperimetric and Analytic Inequalities for Log-Concave Probability Measures , 1999 .

[11]  Prakasa Rao Estimation of a unimodal density , 1969 .

[12]  W. Specht Zur Theorie der elementaren Mittel , 1960 .

[13]  E. Wegman Maximum likelihood estimation of a unimodal density. II , 1970 .

[14]  Sergey G. Bobkov,et al.  The Entropy Per Coordinate of a Random Vector is Highly Constrained Under Convexity Conditions , 2010, IEEE Transactions on Information Theory.

[15]  Alfred O. Hero,et al.  Collaborative 20 Questions for Target Localization , 2013, IEEE Transactions on Information Theory.

[16]  M. Degroot Uncertainty, Information, and Sequential Experiments , 1962 .

[17]  Thomas Kailath,et al.  A coding scheme for additive noise channels with feedback-I: No bandwidth constraint , 1966, IEEE Trans. Inf. Theory.

[18]  D. Bakry,et al.  A simple proof of the Poincaré inequality for a large class of probability measures , 2008 .

[19]  Michael Horstein,et al.  Sequential transmission using noiseless feedback , 1963, IEEE Trans. Inf. Theory.