Partial information decomposition as a unified approach to the specification of neural goal functions

HighlightsUnified framework to compare neural goal functions based on information theory.Domain‐independent description of neural information processing.Enabling principled comparison of neural goal functions.Partial information decomposition for multi‐input systems.Measuring unique, shared and synergistic contributions to neural output. Abstract In many neural systems anatomical motifs are present repeatedly, but despite their structural similarity they can serve very different tasks. A prime example for such a motif is the canonical microcircuit of six‐layered neo‐cortex, which is repeated across cortical areas, and is involved in a number of different tasks (e.g. sensory, cognitive, or motor tasks). This observation has spawned interest in finding a common underlying principle, a ‘goal function’, of information processing implemented in this structure. By definition such a goal function, if universal, cannot be cast in processing‐domain specific language (e.g. ‘edge filtering’, ‘working memory’). Thus, to formulate such a principle, we have to use a domain‐independent framework. Information theory offers such a framework. However, while the classical framework of information theory focuses on the relation between one input and one output (Shannon’s mutual information), we argue that neural information processing crucially depends on the combination of multiple inputs to create the output of a processor. To account for this, we use a very recent extension of Shannon Information theory, called partial information decomposition (PID). PID allows to quantify the information that several inputs provide individually (unique information), redundantly (shared information) or only jointly (synergistic information) about the output. First, we review the framework of PID. Then we apply it to reevaluate and analyze several earlier proposals of information theoretic neural goal functions (predictive coding, infomax and coherent infomax, efficient coding). We find that PID allows to compare these goal functions in a common framework, and also provides a versatile approach to design new goal functions from first principles. Building on this, we design and analyze a novel goal function, called ‘coding with synergy’, which builds on combining external input and prior knowledge in a synergistic manner. We suggest that this novel goal function may be highly useful in neural information processing.

[1]  Tai Sing Lee,et al.  Hierarchical Bayesian inference in the visual cortex. , 2003, Journal of the Optical Society of America. A, Optics, image science, and vision.

[2]  H. B. Barlow,et al.  Possible Principles Underlying the Transformations of Sensory Messages , 2012 .

[3]  M. Larkum A cellular mechanism for cortical associations: an organizing principle for the cerebral cortex , 2013, Trends in Neurosciences.

[4]  Randall D. Beer,et al.  Nonnegative Decomposition of Multivariate Information , 2010, ArXiv.

[5]  Eckehard Olbrich,et al.  Information Decomposition and Synergy , 2015, Entropy.

[6]  M. Sur,et al.  Visual behaviour mediated by retinal projections directed to the auditory pathway , 2000, Nature.

[7]  Jim W Kay,et al.  Coherent Infomax as a Computational Goal for Neural Systems , 2011, Bulletin of mathematical biology.

[8]  Karl J. Friston,et al.  Free-energy and the brain , 2007, Synthese.

[9]  Adam B. Barrett,et al.  An exploration of synergistic and redundant information sharing in static and dynamical Gaussian systems , 2014, Physical review. E, Statistical, nonlinear, and soft matter physics.

[10]  A. Clark,et al.  On the functions, mechanisms, and malfunctions of intracortical contextual modulation , 2015, Neuroscience & Biobehavioral Reviews.

[11]  B J Craven,et al.  Interactions between coincident and orthogonal cues to texture boundaries , 2000, Perception & psychophysics.

[12]  Joseph T. Lizier,et al.  Towards a synergy-based approach to measuring information modification , 2013, 2013 IEEE Symposium on Artificial Life (ALife).

[13]  Jim Kay,et al.  The discovery of structure by multi-stream networks of local processors with contextual guidance , 1995 .

[14]  Raúl Rojas,et al.  Statistics and Neural Networks , 1996 .

[15]  Christoph Salge,et al.  A Bivariate Measure of Redundant Information , 2012, Physical review. E, Statistical, nonlinear, and soft matter physics.

[16]  Ralph Linsker,et al.  Self-organization in a perceptual network , 1988, Computer.

[17]  Schreiber,et al.  Measuring information transfer , 2000, Physical review letters.

[18]  Jim Kay,et al.  Neural networks for unsupervised learning based on information theory , 2000 .

[19]  J. Hohwy The Predictive Mind , 2013 .

[20]  P. König,et al.  A Model of the Ventral Visual System Based on Temporal Stability and Local Memory , 2006, PLoS biology.

[21]  Karl J. Friston,et al.  A free energy principle for the brain , 2006, Journal of Physiology-Paris.

[22]  M. Tribus,et al.  Probability theory: the logic of science , 2003 .

[23]  Ralf Der,et al.  Homeokinesis - A new principle to back up evolution with learning , 1999 .

[24]  Eckehard Olbrich,et al.  Reconsidering unique information: Towards a multivariate information decomposition , 2014, 2014 IEEE International Symposium on Information Theory.

[25]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[26]  Viola Priesemann,et al.  Bits from Brains for Biologically Inspired Computing , 2014, Front. Robot. AI.

[27]  Christof Koch,et al.  Quantifying synergistic mutual information , 2012, ArXiv.

[28]  Randall D. Beer,et al.  Generalized Measures of Information Transfer , 2011, ArXiv.

[29]  Alexander Borst,et al.  One Rule to Grow Them All: A General Theory of Neuronal Branching and Its Practical Application , 2010, PLoS Comput. Biol..

[30]  Niu Guixia A study on college students' capacity for autonomous learning in english based on AHP , 2010, 2010 Second International Conference on Computational Intelligence and Natural Computing.

[31]  Takeshi Kaneko,et al.  Recurrent Infomax Generates Cell Assemblies, Neuronal Avalanches, and Simple Cell-Like Selectivity , 2009, Neural Computation.

[32]  T. Sejnowski,et al.  Brain and cognition , 1989 .

[33]  Albert Y. Zomaya,et al.  Local information transfer as a spatiotemporal filter for complex systems. , 2008, Physical review. E, Statistical, nonlinear, and soft matter physics.

[34]  Eckehard Olbrich,et al.  Quantifying unique information , 2013, Entropy.

[35]  Dario Floreano,et al.  Contextually guided unsupervised learning using local multivariate binary processors , 1998, Neural Networks.

[36]  A. Clark Whatever next? Predictive brains, situated agents, and the future of cognitive science. , 2013, The Behavioral and brain sciences.

[37]  Albert Y. Zomaya,et al.  Local measures of information storage in complex distributed computation , 2012, Inf. Sci..

[38]  Rajesh P. N. Rao,et al.  Predictive coding in the visual cortex: a functional interpretation of some extra-classical receptive-field effects. , 1999 .

[39]  P ? ? ? ? ? ? ? % ? ? ? ? , 1991 .

[40]  G. Zajicek,et al.  The Wisdom of the Body , 1934, Nature.

[41]  Karl J. Friston The free-energy principle: a rough guide to the brain? , 2009, Trends in Cognitive Sciences.

[42]  Hans Rademacher,et al.  The ϑ-functions , 1973 .

[43]  Terrence J. Sejnowski,et al.  The “independent components” of natural scenes are edge filters , 1997, Vision Research.