On the Application of Algorithmic Probability to Autoregressive Models

Typically, the first step in carrying out predictions is to develop an inductive model. In many instances, the best model is used, and it is often selected based on certain relevant criteria that may possibly ignore some of the model uncertainties. The Bayesian approach to model selection uses a weighted average of a class of models thereby overcoming some of the uncertainties associated with selecting the best model. This approach is referred to in the literature as the Bayesian Model Averaging (BMA) approach. It turns out that this approach has significant overlap with the theory of Algorithmic Probability (ALP) developed by R. J. Solomonoff in the early 1960s. The purpose of this article is to first highlight this connection by applying ALP to a set of nested stationary autoregressive time series models, and to give an algorithm to compute “relative weights” of models. This reveals, although empirically, a model weight that can be compared with the Schwarz Bayesian Selection criterion (BIC or SIC). We then develop an elementary algorithm of the Monte Carlo type to evaluate multidimensional integrals over stability domains, and use it to compute what we call the “trimmed weights”.

[1]  Ray J. Solomonoff,et al.  A Formal Theory of Inductive Inference. Part II , 1964, Inf. Control..

[2]  Donald Ervin Knuth,et al.  The Art of Computer Programming , 1968 .

[3]  C. S. Wallace,et al.  An Information Measure for Classification , 1968, Comput. J..

[4]  O. Barndorff-Nielsen,et al.  On the parametrization of autoregressive models by partial autocorrelations , 1973 .

[5]  H. Akaike A new look at the statistical model identification , 1974 .

[6]  J. Makhoul,et al.  Linear prediction: A tutorial review , 1975, Proceedings of the IEEE.

[7]  G. Schwarz Estimating the Dimension of a Model , 1978 .

[8]  J. Rissanen,et al.  Modeling By Shortest Data Description* , 1978, Autom..

[9]  Donald E. Knuth,et al.  The art of computer programming. Vol.2: Seminumerical algorithms , 1981 .

[10]  D. Piccolo THE SIZE OF THE STATIONARITY AND INVERTIBILITY REGION OF AN AUTOREGRESSIVE‐MOVING AVERAGE PROCESS , 1982 .

[11]  J. Monahan A note on enforcing stationarity in autoregressive-moving average models , 1984 .

[12]  S. Shlien,et al.  A geometric description of stable linear predictive coding digital filters , 1985, IEEE Trans. Inf. Theory.

[13]  M. C. Jones Randomly Choosing Parameters from the Stationarity and Invertibility Region of Autoregressive–Moving Average Models , 1987 .

[14]  A. Fam The volume of the coefficient space stability domain of monic polynomials , 1989, IEEE International Symposium on Circuits and Systems,.

[15]  Clifford M. Hurvich,et al.  Regression and time series model selection in small samples , 1989 .

[16]  Ray J. Solomonoff,et al.  The Discovery of Algorithmic Probability , 1997, J. Comput. Syst. Sci..

[17]  David L. Dowe,et al.  Minimum Message Length and Kolmogorov Complexity , 1999, Comput. J..

[18]  Adrian E. Raftery,et al.  Bayesian model averaging: a tutorial (with comments by M. Clyde, David Draper and E. I. George, and a rejoinder by the authors , 1999 .

[19]  Yu. P. Nikolaev The Multidimensional Asymptotic Stability Domain of Linear Discrete Systems: Its Symmetry and Other Properties , 2001 .

[20]  R. Solomonoff A PRELIMINARY REPORT ON A GENERAL THEORY OF INDUCTIVE INFERENCE , 2001 .

[21]  David R. Anderson,et al.  Model selection and multimodel inference : a practical information-theoretic approach , 2003 .

[22]  Feng Liang,et al.  Exact minimax strategies for predictive density estimation, data compression, and model selection , 2002, IEEE Transactions on Information Theory.

[23]  Leigh J. Fitzgibbon,et al.  Minimum message length autoregressive model order selection , 2004, International Conference on Intelligent Sensing and Information Processing, 2004. Proceedings of.

[24]  Marcus Hutter,et al.  Algorithmic Information Theory , 1977, IBM J. Res. Dev..