Pointwise Partial Information DecompositionUsing the Specificity and Ambiguity Lattices

What are the distinct ways in which a set of predictor variables can provide information about a target variable? When does a variable provide unique information, when do variables share redundant information, and when do variables combine synergistically to provide complementary information? The redundancy lattice from the partial information decomposition of Williams and Beer provided a promising glimpse at the answer to these questions. However, this structure was constructed using a much criticised measure of redundant information, and despite sustained research, no completely satisfactory replacement measure has been proposed. In this paper, we take a different approach, applying the axiomatic derivation of the redundancy lattice to a single realisation from a set of discrete variables. To overcome the difficulty associated with signed pointwise mutual information, we apply this decomposition separately to the unsigned entropic components of pointwise mutual information which we refer to as the specificity and ambiguity. This yields a separate redundancy lattice for each component. Then based upon an operational interpretation of redundancy, we define measures of redundant specificity and ambiguity enabling us to evaluate the partial information atoms in each lattice. These atoms can be recombined to yield the sought-after multivariate information decomposition. We apply this framework to canonical examples from the literature and discuss the results and the various properties of the decomposition. In particular, the pointwise decomposition using specificity and ambiguity satisfies a chain rule over target variables, which provides new insights into the so-called two-bit-copy example.

[1]  Robert M. Gray,et al.  Probability, Random Processes, And Ergodic Properties , 1987 .

[2]  Eckehard Olbrich,et al.  On extractable shared information , 2017, Entropy.

[3]  Judea Pearl,et al.  Probabilistic reasoning in intelligent systems - networks of plausible inference , 1991, Morgan Kaufmann series in representation and reasoning.

[4]  Daniel Chicharro,et al.  Synergy and Redundancy in Dual Decompositions of Mutual Information Gain and Information Loss , 2016, Entropy.

[5]  Christoph Salge,et al.  A Bivariate Measure of Redundant Information , 2012, Physical review. E, Statistical, nonlinear, and soft matter physics.

[6]  Marian Verhelst,et al.  Understanding Interdependency Through Complex Information Sharing , 2015, Entropy.

[7]  Nihat Ay,et al.  Hierarchical Quantification of Synergy in Channels , 2016, Front. Robot. AI.

[8]  Angelika Bayer,et al.  A First Course In Probability , 2016 .

[9]  Pinaki Chaudhury,et al.  Information Theoretical Study of Cross-Talk Mediated Signal Transduction in MAPK Pathways , 2015, Entropy.

[10]  Malte Harder,et al.  Information driven self-organization of agents and agent collectives , 2014 .

[11]  P. M. Woodward,et al.  Information theory and inverse probability in telecommunication , 1952 .

[12]  Murray Shanahan,et al.  The Partial Information Decomposition of Generative Neural Network Models , 2017, Entropy.

[13]  Tracey Ho,et al.  Quantifying Redundant Information in Predicting a Target Random Variable , 2014, Entropy.

[14]  John L. Kelly,et al.  A new interpretation of information rate , 1956, IRE Trans. Inf. Theory.

[15]  G. Rota On the foundations of combinatorial theory I. Theory of Möbius Functions , 1964 .

[16]  Daniel Chicharro,et al.  Invariant Components of Synergy, Redundancy, and Unique Information among Three Variables , 2017, Entropy.

[17]  Sang Joon Kim,et al.  A Mathematical Theory of Communication , 2006 .

[18]  Viola Priesemann,et al.  Quantifying Information Modification in Developing Neural Networks via Partial Information Decomposition , 2017, Entropy.

[19]  Dirk Oliver Theis,et al.  Bivariate Partial Information Decomposition: The Optimization Perspective , 2017, Entropy.

[20]  Jim Kay,et al.  Partial and Entropic Information Decompositions of a Neuronal Modulatory Interaction , 2017, Entropy.

[21]  Luca Faes,et al.  Synergetic and redundant information flow detected by unnormalized Granger causality: application to resting state fMRI. , 2015, IEEE transactions on bio-medical engineering.

[22]  Randall D. Beer,et al.  Nonnegative Decomposition of Multivariate Information , 2010, ArXiv.

[23]  Luca Faes,et al.  Multiscale Information Decomposition: Exact Computation for Multivariate Gaussian Processes , 2017, Entropy.

[24]  Joseph T. Lizier,et al.  Probability Mass Exclusions and the Directed Components of Pointwise Mutual Information , 2018, ArXiv.

[25]  Robin A. A. Ince The Partial Entropy Decomposition: Decomposing multivariate entropy and mutual information via pointwise common surprisal , 2017, ArXiv.

[26]  Eckehard Olbrich,et al.  Quantifying unique information , 2013, Entropy.

[27]  Eckehard Olbrich,et al.  Reconsidering unique information: Towards a multivariate information decomposition , 2014, 2014 IEEE International Symposium on Information Theory.

[28]  Daniele Marinazzo,et al.  Redundant variables and Granger causality. , 2009, Physical review. E, Statistical, nonlinear, and soft matter physics.

[29]  G. Rota On the Foundations of Combinatorial Theory , 2009 .

[30]  Mill Johannes G.A. Van,et al.  Transmission Of Information , 1961 .

[31]  James P. Crutchfield,et al.  Multivariate Dependence Beyond Shannon Information , 2016, Entropy.

[32]  N. Martin,et al.  Mathematical Theory of Entropy , 1981 .

[33]  Eckehard Olbrich,et al.  Shared Information -- New Insights and Problems in Decomposing Information in Complex Systems , 2012, ArXiv.

[34]  Keyan Zahedi,et al.  Morphological Computation: Synergy of Body and Brain , 2017, Entropy.

[35]  James P. Crutchfield,et al.  Intersection Information Based on Common Randomness , 2013, Entropy.

[36]  David J. C. MacKay,et al.  Information Theory, Inference, and Learning Algorithms , 2004, IEEE Transactions on Information Theory.

[37]  Robin A. A. Ince Measuring multivariate redundant information with pointwise common change in surprisal , 2016, Entropy.

[38]  Philip M. Woodward,et al.  Probability and Information Theory with Applications to Radar , 1954 .

[39]  Eckehard Olbrich,et al.  Information Decomposition and Synergy , 2015, Entropy.

[40]  Christof Koch,et al.  Quantifying synergistic mutual information , 2012, ArXiv.

[41]  Charalambos A. Charalambides,et al.  Enumerative combinatorics , 2018, SIGA.

[42]  Adam B. Barrett,et al.  An exploration of synergistic and redundant information sharing in static and dynamical Gaussian systems , 2014, Physical review. E, Statistical, nonlinear, and soft matter physics.

[43]  Eckehard Olbrich,et al.  Coarse-Graining and the Blackwell Order , 2017, Entropy.

[44]  Joseph T. Lizier,et al.  Towards a synergy-based approach to measuring information modification , 2013, 2013 IEEE Symposium on Artificial Life (ALife).

[45]  Nathaniel Virgo,et al.  Decomposing multivariate information , 2017 .

[46]  Joseph T. Lizier,et al.  Probability Mass Exclusions and the Directed Components of Mutual Information , 2018, Entropy.

[47]  Brian A. Davey,et al.  An Introduction to Lattices and Order , 1989 .

[48]  Johannes Rauh,et al.  Secret Sharing and Shared Information , 2017, Entropy.