Estimations of Integrated Information Based on Algorithmic Complexity and Dynamic Querying

The concept of information has emerged as a language in its own right, bridging several disciplines that analyze natural phenomena and man-made systems. Integrated information has been introduced as a metric to quantify the amount of information generated by a system beyond the information generated by its elements. Yet, this intriguing notion comes with the price of being prohibitively expensive to calculate, since the calculations require an exponential number of sub-divisions of a system. Here we introduce a novel framework to connect algorithmic randomness and integrated information and a numerical method for estimating integrated information using a perturbation test rooted in algorithmic information dynamics. This method quantifies the change in program size of a system when subjected to a perturbation. The intuition behind is that if an object is random then random perturbations have little to no effect to what happens when a shorter program but when an object has the ability to move in both directions (towards or away from randomness) it will be shown to be better integrated as a measure of sophistication telling apart randomness and simplicity from structure. We show that an object with a high integrated information value is also more compressible, and is, therefore, more sensitive to perturbations. We find that such a perturbation test quantifying compression sensitivity provides a system with a means to extract explanations--causal accounts--of its own behaviour. Our technique can reduce the number of calculations to arrive at some bounds or estimations, as the algorithmic perturbation test guides an efficient search for estimating integrated information. Our work sets the stage for a systematic exploration of connections between algorithmic complexity and integrated information at the level of both theory and practice.

[1]  P. Anderson More is different. , 1972, Science.

[2]  Jean-Paul Delahaye,et al.  Numerical evaluation of algorithmic complexity for short strings: A glance into the innermost structure of randomness , 2011, Appl. Math. Comput..

[3]  Hector Zenil,et al.  Algorithmic Information Dynamics of Emergent, Persistent, and Colliding Particles in the Game of Life , 2019, From Parallel to Emergent Computing.

[4]  Schreiber,et al.  Measuring information transfer , 2000, Physical review letters.

[5]  Giulio Ruffini,et al.  An algorithmic information theory of consciousness , 2017, Neuroscience of consciousness.

[6]  Larissa Albantakis,et al.  PyPhi: A toolbox for integrated information theory , 2017, PLoS Comput. Biol..

[7]  Hector Zenil,et al.  A Review of Graph and Network Complexity from an Algorithmic Information Perspective , 2018, Entropy.

[8]  Larissa Albantakis,et al.  From the Phenomenology to the Mechanisms of Consciousness: Integrated Information Theory 3.0 , 2014, PLoS Comput. Biol..

[9]  Hector Zenil,et al.  A Decomposition Method for Global Evaluation of Shannon Entropy and Local Estimations of Algorithmic Complexity , 2016, Entropy.

[10]  C. Granger Investigating causal relations by econometric models and cross-spectral methods , 1969 .

[11]  Peter Dayan,et al.  Theoretical Neuroscience: Computational and Mathematical Modeling of Neural Systems , 2001 .

[12]  James P. Crutchfield,et al.  Computational Mechanics: Pattern and Prediction, Structure and Simplicity , 1999, ArXiv.

[13]  Victor Iapascurta,et al.  Detection of Movement toward Randomness by Applying the Block Decomposition Method to a Simple Model of the Circulatory System , 2019, Complex Syst..

[14]  Hector Zenil,et al.  An Algorithmic Information Calculus for Causal Discovery and Reprogramming Systems , 2017, bioRxiv.

[15]  Hector Zenil,et al.  Methods of information theory and algorithmic complexity for network biology. , 2014, Seminars in cell & developmental biology.

[16]  C. Koch,et al.  Integrated information theory: from consciousness to its physical substrate , 2016, Nature Reviews Neuroscience.

[17]  Albert-Lszl Barabsi,et al.  Network Science , 2016, Encyclopedia of Big Data.

[18]  Gregory J. Chaitin,et al.  On the Length of Programs for Computing Finite Binary Sequences , 1966, JACM.

[19]  C. Villani Topics in Optimal Transportation , 2003 .

[20]  G. Tononi,et al.  A Theoretically Based Index of Consciousness Independent of Sensory Processing and Behavior , 2013, Science Translational Medicine.

[21]  Hector Zenil,et al.  Coding-theorem like behaviour and emergence of the universal distribution from resource-bounded algorithmic probability , 2017, Int. J. Parallel Emergent Distributed Syst..

[22]  Steven H. Strogatz,et al.  Nonlinear Dynamics And Chaos With Applications To Physics Biology Chemistry And Engineering Studies In Nonlinearity , 2021 .

[23]  Hector Zenil,et al.  Algorithmic Data Analytics, Small Data Matters and Correlation versus Causation , 2013, 1309.1418.

[24]  J. Crutchfield,et al.  Thermodynamic depth of causal states: Objective complexity via minimal representations , 1999 .

[25]  Hector Zenil,et al.  Low Algorithmic Complexity Entropy-deceiving Graphs , 2016, Physical review. E.

[26]  Hector Zenil,et al.  A Behavioural Foundation for Natural Computing and a Programmability Test , 2013, ArXiv.

[27]  Hector Zenil,et al.  A Computable Universe: Understanding and Exploring Nature As Computation , 2012 .

[28]  Mario Ventresca,et al.  Using Algorithmic Complexity to Differentiate Cognitive States in fMRI , 2018, COMPLEX NETWORKS.

[29]  Steven H. Strogatz,et al.  Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering , 1994 .

[30]  J. Pearl Causality: Models, Reasoning and Inference , 2000 .

[31]  John H. Holland,et al.  Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence , 1992 .

[32]  Jean-Paul Delahaye,et al.  Calculating Kolmogorov Complexity from the Output Frequency Distributions of Small Turing Machines , 2012, PloS one.

[33]  Hector Zenil,et al.  Testing Biological Models for Non-linear Sensitivity with a Programmability Test , 2013, ECAL.

[34]  David Bailey,et al.  On the rapid computation of various polylogarithmic constants , 1997, Math. Comput..

[35]  Angelika Schmidt,et al.  Causality, Information and Biological Computation: An algorithmic software approach to life, disease and the immune system , 2015, 1508.06538.

[36]  Ulrik Brandes,et al.  What is network science? , 2013, Network Science.

[37]  A. Kolmogorov Three approaches to the quantitative definition of information , 1968 .

[38]  Ryota Kanai,et al.  Efficient Algorithms for Searching the Minimum Information Partition in Integrated Information Theory , 2017, Entropy.

[39]  Sean Devine The application of algorithmic information theory to noisy patterned strings , 2006, Complex..

[40]  Pierre Vandergheynst,et al.  Geometric Deep Learning: Going beyond Euclidean data , 2016, IEEE Signal Process. Mag..

[41]  Hector Zenil,et al.  Correlation of automorphism group size and topological properties with program−size complexity evaluations of graphs and complex networks , 2013, 1306.0322.

[42]  Cosma Rohilla Shalizi Optimal Nonlinear Prediction of Random Fields on Networks , 2003, DMCS.