A Kullback-Leibler View of Maximum Entropy and Maximum Log-Probability Methods

Entropy methods enable a convenient general approach to providing a probability distribution with partial information. The minimum cross-entropy principle selects the distribution that minimizes the Kullback–Leibler divergence subject to the given constraints. This general principle encompasses a wide variety of distributions, and generalizes other methods that have been proposed independently. There remains, however, some confusion about the breadth of entropy methods in the literature. In particular, the asymmetry of the Kullback–Leibler divergence provides two important special cases when the target distribution is uniform: the maximum entropy method and the maximum log-probability method. This paper compares the performance of both methods under a variety of conditions. We also examine a generalized maximum log-probability method as a further demonstration of the generality of the entropy approach.

[1]  J. Eric Bickel,et al.  Approximating Joint Probability Distributions Given Partial Information , 2013, Decis. Anal..

[2]  Edwin T. Jaynes,et al.  Prior Probabilities , 1968, Encyclopedia of Machine Learning.

[3]  Ali E. Abbas,et al.  Maximum Entropy Distributions between Upper and Lower Bounds , 2005 .

[4]  Ali E. Abbas,et al.  An Entropy Approach for Utility Assignment in Decision Analysis , 2003 .

[5]  Ali E. Abbas,et al.  Entropy methods for adaptive utility elicitation , 2004, IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans.

[6]  Yvonne Freeh,et al.  Interior Point Algorithms Theory And Analysis , 2016 .

[7]  Ali E. Abbas,et al.  Entropy methods for joint distributions in decision analysis , 2006, IEEE Transactions on Engineering Management.

[8]  Claude E. Shannon,et al.  Communication theory of secrecy systems , 1949, Bell Syst. Tech. J..

[9]  G. Sonnevend An "analytical centre" for polyhedrons and new classes of global algorithms for linear (smooth, convex) programming , 1986 .

[10]  G. C. Tiao,et al.  Bayesian inference in statistical analysis , 1973 .

[11]  Ali E. Abbas,et al.  A Kullback-Leibler View of Linear and Log-Linear Pools , 2009, Decis. Anal..

[12]  G. Sonnevend,et al.  Applications of the notion of analytic center in approximation (estimation) problems , 1989 .

[13]  R. Nelsen An Introduction to Copulas , 1998 .

[14]  Stephen P. Boyd,et al.  Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.

[15]  Dirk P. Kroese,et al.  The Cross-Entropy Method: A Unified Approach to Combinatorial Optimization, Monte-Carlo Simulation and Machine Learning , 2004 .

[16]  R. A. Leibler,et al.  On Information and Sufficiency , 1951 .

[17]  Ali E. Abbas,et al.  Entropy Methods For Univariate Distributions in Decision Analysis , 2003 .

[18]  P. Fishburn Foundations of decision analysis along the way , 1989 .

[19]  E. Jaynes Information Theory and Statistical Mechanics , 1957 .

[20]  S. Kullback,et al.  Information Theory and Statistics , 1959 .