Algorithmic statistics

While Kolmogorov (1965, 1983) complexity is the accepted absolute measure of information content of an individual finite object, a similarly absolute notion is needed for the relation between an individual data sample and an individual model summarizing the information in the data, for example, a finite set (or probability distribution) where the data sample typically came from. The statistical theory based on such relations between individual objects can be called algorithmic statistics, in contrast to classical statistical theory that deals with relations between probabilistic ensembles. We develop the algorithmic theory of statistic, sufficient statistic, and minimal sufficient statistic. This theory is based on two-part codes consisting of the code for the statistic (the model summarizing the regularity, the meaningful information, in the data) and the model-to-data code. In contrast to the situation in probabilistic statistical theory, the algorithmic relation of (minimal) sufficiency is an absolute relation between the individual model and the individual data sample. We distinguish implicit and explicit descriptions of the models. We give characterizations of algorithmic (Kolmogorov) minimal sufficient statistic for all data samples for both description modes-in the explicit mode under some constraints. We also strengthen and elaborate on earlier results for the "Kolmogorov structure function" and "absolutely nonstochastic objects"-those objects for which the simplest models that summarize their relevant information (minimal sufficient statistics) are at least as complex as the objects themselves. We demonstrate a close relation between the probabilistic notions and the algorithmic ones: (i) in both cases there is an "information non-increase" law; (ii) it is shown that a function is a probabilistic sufficient statistic iff it is with high probability (in an appropriate sense) an algorithmic sufficient statistic.

[1]  Péter Gács,et al.  Towards an Algorithmic Statistics , 2000, ALT.

[2]  J. Lukasiewicz Logical foundations of probability theory , 1970 .

[3]  Ming Li,et al.  An Introduction to Kolmogorov Complexity and Its Applications , 2019, Texts in Computer Science.

[4]  T. Cover Kolmogorov Complexity, Data Compression, and Inference , 1985 .

[5]  Ming Li,et al.  Minimum description length induction, Bayesianism, and Kolmogorov complexity , 1999, IEEE Trans. Inf. Theory.

[6]  Per Martin-Löf,et al.  The Definition of Random Sequences , 1966, Inf. Control..

[7]  Leonid A. Levin,et al.  Randomness Conservation Inequalities; Information and Independence in Mathematical Theories , 1984, Inf. Control..

[8]  V. V'yugin On the Defect of Randomness of a Finite Object with Respect to Measures with Given Complexity Bounds , 1988 .

[9]  R. Fisher,et al.  On the Mathematical Foundations of Theoretical Statistics , 1922 .

[10]  A. Kolmogorov,et al.  ALGORITHMS AND RANDOMNESS , 1988 .

[11]  Gregory J. Chaitin,et al.  A recent technical report , 1974, SIGA.

[12]  Vladimir V. V'yugin,et al.  Algorithmic Complexity and Stochastic Properties of Finite Binary Sequences , 1999, Comput. J..

[13]  Ming Li,et al.  Applying MDL to learn best model granularity , 2000, Artif. Intell..

[14]  J. K. Skwirzynski The impact of processing techniques on communications , 1985 .

[15]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[16]  Alexander Shen,et al.  Discussion on Kolmogorov Complexity and Statistical Analysis , 1999, Comput. J..

[17]  A. Kolmogorov Three approaches to the quantitative definition of information , 1968 .

[18]  A. Kolmogorov On Logical Foundations of Probability Theory , 1983 .

[19]  Jorma Rissanen,et al.  The Minimum Description Length Principle in Coding and Modeling , 1998, IEEE Trans. Inf. Theory.

[20]  Ray J. Solomonoff,et al.  A Formal Theory of Inductive Inference. Part I , 1964, Inf. Control..