Book Inequalities

Information theoretical inequalities have strong ties with polymatroids and their representability. A polymatroid is entropic if its rank function is given by the Shannon entropy of the subsets of some discrete random variables. The book is a special iterated adhesive extension of a polymatroid with the property that entropic polymatroids have n-page book extensions over an arbitrary spine. We prove that every polymatroid has an n-page book extension over a single element and over an all-but-one-element spine. Consequently, for polymatroids on four elements, only book extensions over a two-element spine should be considered. Matúš proved that the Zhang-Yeung inequalities characterize polymatroids on four elements which have such a two-page book extension. The n-page book inequalities, defined in this paper, are conjectured to characterize polymatroids on four elements which have n-page book extensions over a two-element spine. We prove that the condition is necessary; consequently, every book inequality is an information inequality on four random variables. Using computer-aided multiobjective optimization, the sufficiency of the condition is verified up to nine-page book extensions.

[1]  Zhen Zhang,et al.  On Characterization of Entropy Function via Information Inequalities , 1998, IEEE Trans. Inf. Theory.

[2]  László Lovász,et al.  Submodular functions and convexity , 1982, ISMP.

[3]  Frantisek Matús,et al.  Adhesivity of polymatroids , 2007, Discret. Math..

[4]  Lizhen Shao,et al.  A dual variant of Benson’s “outer approximation algorithm” for multiple objective linear programming , 2012, J. Glob. Optim..

[5]  F. Matús,et al.  Two Constructions on Limits of Entropy Functions , 2007, IEEE Transactions on Information Theory.

[6]  S. Weber,et al.  Relationships among bounds for the region of entropic vectors in four variables , 2010, 2010 48th Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[7]  Steven P. Weber,et al.  Matroid bounds on the region of entropic vectors , 2013, 2013 51st Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[8]  Terence Chan Recent Progresses in Characterising Information Inequalities , 2011, Entropy.

[9]  Frantisek Matús,et al.  Infinitely Many Information Inequalities , 2007, 2007 IEEE International Symposium on Information Theory.

[10]  Nikolai K. Vereshchagin,et al.  A new class of non-Shannon-type inequalities for entropies , 2002, Commun. Inf. Syst..

[11]  László Csirmaz,et al.  Entropy Region and Convolution , 2016, IEEE Transactions on Information Theory.

[12]  Randall Dougherty,et al.  Six New Non-Shannon Information Inequalities , 2006, 2006 IEEE International Symposium on Information Theory.

[13]  Weidong Xu,et al.  A projection method for derivation of non-Shannon-type information inequalities , 2008, 2008 IEEE International Symposium on Information Theory.

[14]  Tarik Kaced Equivalence of two proof techniques for non-shannon-type inequalities , 2013, 2013 IEEE International Symposium on Information Theory.

[16]  Harold P. Benson,et al.  An Outer Approximation Algorithm for Generating All Efficient Extreme Points in the Outcome Set of a Multiple Objective Linear Programming Problem , 1998, J. Glob. Optim..

[17]  A. Ingleton,et al.  Conditions for representability and transversality of matroids , 1971 .

[18]  Randall Dougherty,et al.  Non-Shannon Information Inequalities in Four Random Variables , 2011, ArXiv.

[19]  T. H. Chan,et al.  Balanced information inequalities , 2003, IEEE Trans. Inf. Theory.

[20]  G. G. Stokes "J." , 1890, The New Yale Book of Quotations.