The Speed Prior: A New Simplicity Measure Yielding Near-Optimal Computable Predictions

Solomonoff's optimal but noncomputable method for inductive inference assumes that observation sequences x are drawn from an recursive prior distribution µ(x). Instead of using the unknown µ(x) he predicts using the celebrated universal enumerable prior M(x) which for all x exceeds any recursive µ(x), save for a constant factor independent of x. The simplicity measure M(x) naturally implements "Occam's razor" and is closely related to the Kolmogorov complexity of x. However, M assigns high probability to certain data x that are extremely hard to compute. This does not match our intuitive notion of simplicity. Here we suggest a more plausible measure derived from the fastest way of computing data. In absence of contrarian evidence, we assume that the physical world is generated by a computational process, and that any possibly infinite sequence of observations is therefore computable in the limit (this assumption is more radical and stronger than Solomonoff's). Then we replace M by the novel Speed Prior S, under which the cumulative a priori probability of all data whose computation through an optimal algorithm requires more than O(n) resources is 1/n. We show that the Speed Prior allows for deriving a computable strategy for optimal prediction of future y, given past x. Then we consider the case that the data actually stem from a nonoptimal, unknown computational process, and use Hutter's recent results to derive excellent expected loss bounds for S-based inductive inference. We conclude with several nontraditional predictions concerning the future of our universe.

[1]  Leon Gordon Kraft,et al.  A device for quantizing, grouping, and coding amplitude-modulated pulses , 1949 .

[2]  H. Everett "Relative State" Formulation of Quantum Mechanics , 1957 .

[3]  Ray J. Solomonoff,et al.  A Formal Theory of Inductive Inference. Part I , 1964, Inf. Control..

[4]  Ray J. Solomonoff,et al.  A Formal Theory of Inductive Inference. Part II , 1964, Inf. Control..

[5]  A. Kolmogorov Three approaches to the quantitative definition of information , 1968 .

[6]  C. S. Wallace,et al.  An Information Measure for Classification , 1968, Comput. J..

[7]  Gregory J. Chaitin,et al.  On the Length of Programs for Computing Finite Binary Sequences: statistical considerations , 1969, JACM.

[8]  L. Levin,et al.  THE COMPLEXITY OF FINITE OBJECTS AND THE DEVELOPMENT OF THE CONCEPTS OF INFORMATION AND RANDOMNESS BY MEANS OF THE THEORY OF ALGORITHMS , 1970 .

[9]  B. Carter Large Number Coincidences and the Anthropic Principle in Cosmology , 1974 .

[10]  Ray J. Solomonoff,et al.  Complexity-based induction systems: Comparisons and convergence theorems , 1978, IEEE Trans. Inf. Theory.

[11]  Péter Gács On the relation between descriptional complexity and algorithmic probability , 1981, 22nd Annual Symposium on Foundations of Computer Science (sfcs 1981).

[12]  Hans J. Bremermann,et al.  Minimum energy requirements of information transfer and computing , 1982 .

[13]  J. Rissanen Stochastic Complexity and Modeling , 1986 .

[14]  Konrad Zuse,et al.  Rechnender Raum , 1991, Physik und Informatik.

[15]  Ming Li,et al.  An Introduction to Kolmogorov Complexity and Its Applications , 2019, Texts in Computer Science.

[16]  Jürgen Schmidhuber Discovering Solutions with Low Kolmogorov Complexity and High Generalization Capability , 1995, ICML.

[17]  C. H. Bennett,et al.  Quantum Information and Computation , 1995 .

[18]  Corso Elvezia,et al.  Discovering Neural Nets with Low Kolmogorov Complexity and High Generalization Capability , 1997 .

[19]  William I. Gasarch,et al.  Book Review: An introduction to Kolmogorov Complexity and its Applications Second Edition, 1997 by Ming Li and Paul Vitanyi (Springer (Graduate Text Series)) , 1997, SIGACT News.

[20]  Wilfried Brauer,et al.  Foundations of computer science : potential--theory--cognition , 1997 .

[21]  J. Urgen Schmidhuber A Computer Scientist's View of Life, the Universe, and Everything , 1997 .

[22]  G. Hooft Quantum gravity as a dissipative deterministic system , 1999, gr-qc/9903084.

[23]  C. Monroe,et al.  Experimental entanglement of four particles , 2000, Nature.

[24]  S. Lloyd Ultimate physical limits to computation , 1999, Nature.

[25]  C. Schmidhuber Strings from Logic , 2000, hep-th/0011065.

[26]  Jürgen Schmidhuber,et al.  Algorithmic Theories of Everything , 2000, ArXiv.

[27]  Marcus Hutter General Loss Bounds for Universal Sequence Prediction , 2001, ICML.

[28]  Marcus Hutter Convergence and Error Bounds for Universal Prediction of Nonbinary Sequences , 2001, ECML.

[29]  Marcus Hutter,et al.  Towards a Universal Theory of Artificial Intelligence Based on Algorithmic Probability and Sequential Decisions , 2000, ECML.

[30]  Jürgen Schmidhuber,et al.  Hierarchies of Generalized Kolmogorov Complexities and Nonenumerable Universal Measures Computable in the Limit , 2002, Int. J. Found. Comput. Sci..

[31]  Marcus Hutter The Fastest and Shortest Algorithm for all Well-Defined Problems , 2002, Int. J. Found. Comput. Sci..