Testing nonlinear Markovian hypotheses in dynamical systems

Abstract We present a statistical approach for detecting the Markovian character of dynamical systems by analyzing their flow of information. Especially in the presence of noise which is mostly the case for real-world time series, the calculation of the information flow of the underlying system via the concept of symbolic dynamics is rather problematic since one has to use infinitesimal partitions. We circumvent this difficulty by measuring the information flow indirectly. More precisely, we calculate a measure based on higher order cumulants which quantifies the statistical dependencies between the past values of the time series and the point r steps ahead. As an extension of Theiler's method of surrogate data (Theiler et al., 1992) this cumulant based information flow (a function of the look-ahead r ) is used as the discriminating statistic in testing the observed dynamics against a hierarchy of null hypotheses corresponding to nonlinear Markov processes of increasing order. This procedure is iterative in the sense that whenever a null hypothesis is rejected new data sets can be generated corresponding to better approximations of the original process in terms of information flow. Since we use higher order cumulants for calculating the discriminating statistic our method is also applicable to small data sets. Numerical results on artificial and real-world examples including non-chaotic, nonlinear processes, autoregressive models and noisy chaos show the effectiveness of our approach.

[1]  D. Ruelle,et al.  Ergodic theory of chaos and strange attractors , 1985 .

[2]  A. Wolf,et al.  Determining Lyapunov exponents from a time series , 1985 .

[3]  E. Lorenz Deterministic nonperiodic flow , 1963 .

[4]  Ralph Neuneier,et al.  Estimation of Conditional Densities: A Comparison of Neural Network Approaches , 1994 .

[5]  G. Deco,et al.  An Information-Theoretic Approach to Neural Computing , 1997, Perspectives in Neural Computing.

[6]  Deco,et al.  Learning time series evolution by unsupervised extraction of correlations. , 1995, Physical review. E, Statistical physics, plasmas, fluids, and related interdisciplinary topics.

[7]  Christopher M. Bishop,et al.  Neural networks for pattern recognition , 1995 .

[8]  Bernd Pompe,et al.  Chaos in dissipativen Systemen , 1989 .

[9]  Albano,et al.  Kolmogorov-Smirnov test distinguishes attractors with similar dimensions. , 1995, Physical review. E, Statistical physics, plasmas, fluids, and related interdisciplinary topics.

[10]  M. Rosenblatt Remarks on Some Nonparametric Estimates of a Density Function , 1956 .

[11]  Ichiro Tsuda,et al.  Information theoretical approach to noisy dynamics , 1985 .

[12]  Györgyi,et al.  Calculation of the entropy in chaotic systems. , 1985, Physical review. A, General physics.

[13]  Athanasios Papoulis,et al.  Probability, Random Variables and Stochastic Processes , 1965 .

[14]  E. Parzen On Estimation of a Probability Density Function and Mode , 1962 .

[15]  C. Gardiner Handbook of Stochastic Methods , 1983 .

[16]  H. Tong Non-linear time series. A dynamical system approach , 1990 .

[17]  Deco,et al.  Statistical-ensemble theory of redundancy reduction and the duality between unsupervised and supervised neural learning. , 1995, Physical review. E, Statistical physics, plasmas, fluids, and related interdisciplinary topics.

[18]  Györgyi,et al.  Entropy decay as a measure of stochasticity in chaotic systems. , 1986, Physical review. A, General physics.

[19]  James Theiler,et al.  Testing for nonlinearity in time series: the method of surrogate data , 1992 .

[20]  P. Grassberger,et al.  Characterization of Strange Attractors , 1983 .

[21]  B. Silverman Density estimation for statistics and data analysis , 1986 .

[22]  L. Breiman Statistics: with a view toward applications , 1969 .

[23]  N. Packard,et al.  Symbolic dynamics of noisy chaos , 1983 .

[24]  G. Deco,et al.  Exploring the intrinsic information loss in single-humped maps by refining multi-symbol partitions , 1996 .

[25]  Andrew M. Fraser,et al.  Information and entropy in strange attractors , 1989, IEEE Trans. Inf. Theory.