On the Ingleton-Violations in Finite Groups

Given $n$ discrete random variables, its entropy vector is the $2^n-1$ dimensional vector obtained from the joint entropies of all non-empty subsets of the random variables. It is well known that there is a one-to-one correspondence between such an entropy vector and a certain group-characterizable vector obtained from a finite group and $n$ of its subgroups [3]. This correspondence may be useful for characterizing the space of entropic vectors and for designing network codes. If one restricts attention to abelian groups then not all entropy vectors can be obtained. This is an explanation for the fact shown by Dougherty et al [4] that linear network codes cannot achieve capacity in general network coding problems. All abelian group-characterizable vectors, and by fiat all entropy vectors generated by linear network codes, satisfy a linear inequality called the Ingleton inequality. It is therefore of interest to identify groups that violate the Ingleton inequality. In this paper, we study the problem of finding nonabelian finite groups that yield characterizable vectors which violate the Ingleton inequality. Using a refined computer search, we find the symmetric group $S_5$ to be the smallest group that violates the Ingleton inequality. Careful study of the structure of this group, and its subgroups, reveals that it belongs to the Ingleton-violating family $PGL(2,q)$ with a prime power $q \geq 5$, i.e., the projective group of $2\times 2$ nonsingular matrices with entries in $\mathbb{F}_q$. We further interpret this family using the theory of group actions. We also extend the construction to more general groups such as $PGL(n,q)$ and $GL(n,q)$. The families of groups identified here are therefore good candidates for constructing network codes more powerful than linear network codes, and we discuss some considerations for constructing such group network codes.

[1]  Terence Chan,et al.  Group characterizable entropy functions , 2007, 2007 IEEE International Symposium on Information Theory.

[2]  Terence Chan On the optimality of group network codes , 2005, Proceedings. International Symposium on Information Theory, 2005. ISIT 2005..

[3]  Babak Hassibi,et al.  Normalized Entropy Vectors, Network Information Theory and Convex Optimization , 2007, 2007 IEEE Information Theory Workshop on Information Theory for Wireless Networks.

[4]  Raymond W. Yeung,et al.  On a relation between information inequalities and group theory , 2002, IEEE Trans. Inf. Theory.

[5]  Ryan Kinser,et al.  New inequalities for subspace arrangements , 2009, J. Comb. Theory, Ser. A.

[6]  Zhen Zhang,et al.  A non-Shannon-type conditional inequality of information quantities , 1997, IEEE Trans. Inf. Theory.

[7]  Babak Hassibi,et al.  On group network codes: Ingleton-bound violations and independent sources , 2010, 2010 IEEE International Symposium on Information Theory.

[8]  František Matúš,et al.  Conditional Independences among Four Random Variables III: Final Conclusion , 1999, Combinatorics, probability & computing.

[9]  Randall Dougherty,et al.  Linear rank inequalities on five or more variables , 2009, ArXiv.

[10]  Randall Dougherty,et al.  Non-Shannon Information Inequalities in Four Random Variables , 2011, ArXiv.

[11]  Randall Dougherty,et al.  Networks, Matroids, and Non-Shannon Information Inequalities , 2007, IEEE Transactions on Information Theory.

[12]  M. Lunelli,et al.  Representation of matroids , 2002, math/0202294.

[13]  Nikolai K. Vereshchagin,et al.  Inequalities for Shannon Entropy and Kolmogorov Complexity , 1997, J. Comput. Syst. Sci..

[14]  Frédérique E. Oggier,et al.  Groups and information inequalities in 5 variables , 2013, 2013 51st Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[15]  Alex J. Grant,et al.  Truncation Technique for Characterizing Linear Polymatroids , 2011, IEEE Transactions on Information Theory.

[16]  Randall Dougherty,et al.  Insufficiency of linear coding in network information flow , 2005, IEEE Transactions on Information Theory.

[17]  Babak Hassibi,et al.  MCMC methods for entropy optimization and nonlinear network coding , 2010, 2010 IEEE International Symposium on Information Theory.

[18]  B. Hassibi,et al.  Cayley's hyperdeterminant, the principal minors of a symmetric matrix and the entropy region of 4 Gaussian random variables , 2008, 2008 46th Annual Allerton Conference on Communication, Control, and Computing.

[19]  Pirita Paajanen,et al.  Finite p-Groups, Entropy Vectors, and the Ingleton Inequality for Nilpotent Groups , 2013, IEEE Transactions on Information Theory.

[20]  Zhen Zhang,et al.  On Characterization of Entropy Function via Information Inequalities , 1998, IEEE Trans. Inf. Theory.

[21]  Hua Li,et al.  On Connections between Group Homomorphisms and the Ingleton Inequality , 2007, 2007 IEEE International Symposium on Information Theory.

[22]  Frantisek Matús,et al.  Conditional Independences among Four Random Variables II , 1995, Combinatorics, Probability and Computing.

[23]  T. Chan,et al.  Capacity regions for linear and abelian network codes , 2007, 2007 Information Theory and Applications Workshop.

[24]  Nigel Boston,et al.  Large violations of the Ingleton inequality , 2012, 2012 50th Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[25]  S. Shadbakht Entropy Region and Network Information Theory , 2011 .