Twofold universal prediction schemes for achieving the finite-state predictability of a noisy individual binary sequence

The problem of predicting the next outcome of an individual binary sequence corrupted by noise using finite memory, is considered. The conditional finite-state (FS) predictability of an infinite individual sequence given its noisy version is defined as the minimum fraction of errors that can be made by any FS predictor fed by the noisy version. It is proved that the conditional FS predictability can be attained almost surely by universal sequential prediction schemes in the case where the noisy version is the output of a binary-symmetric channel (BSC) whose input is the clean individual sequence. In particular, universal predictors of the original noise-free setting, which operate on the noisy sequence, have this property. Moreover, these universal predictors do not depend on the crossover probability characterizing the BSC. It is seen that the noisy setting gives rise to additional criteria by which the performance of prediction schemes can be assessed. Finally, a closer look is taken at the conditional FS predictability, and this quantity is proposed as an additional measure of the complexity of a sequence, perhaps finer and more informative than the predictive complexity of the noise-free setting.

[1]  H. Robbins,et al.  Asymptotic Solutions of the Compound Decision Problem for Two Completely Specified Distributions , 1955 .

[2]  H. Robbins,et al.  TESTING STATISTICAL HYPOTHESES. THE COMPOUND APPROACH , 1961 .

[3]  R. Rao Relations between Weak and Uniform Convergence of Measures with Applications , 1962 .

[4]  E. Samuel An Empirical Bayes Approach to the Testing of Certain Parametric Hypotheses , 1963 .

[5]  E. Samuel Asymptotic Solutions of the Sequential Compound Decision Problem , 1963 .

[6]  E. Samuel NOTE ON A SEQUENTIAL CLASSIFICATION PROBLEM , 1963 .

[7]  E. Samuel Convergence of the Losses of Certain Decision Rules for the Sequential Compound Decision Problem , 1964 .

[8]  Ray J. Solomonoff,et al.  A Formal Theory of Inductive Inference. Part I , 1964, Inf. Control..

[9]  Ray J. Solomonoff,et al.  A Formal Theory of Inductive Inference. Part II , 1964, Inf. Control..

[10]  E. Samuel On Simple Rules for the Compound Decision Problem , 1965 .

[11]  E. Samuel Sequential Compound Estimators , 1965 .

[12]  E. Samuel The compound decision problem in the opponent case , 1965 .

[13]  J. V. Ryzin,et al.  The Sequential Compound Decision Problem with $m \times n$ Finite Loss Matrix , 1966 .

[14]  J. V. Ryzin,et al.  The Compound Decision Problem with $m \times n$ Finite Loss Matrix , 1966 .

[15]  A. Kolmogorov Three approaches to the quantitative definition of information , 1968 .

[16]  W. Stout The Hartman-Wintner Law of the Iterated Logarithm for Martingales , 1970 .

[17]  W. Stout A martingale analogue of Kolmogorov's law of the iterated logarithm , 1970 .

[18]  Aaron D. Wyner,et al.  A theorem on the entropy of certain binary sequences and applications-I , 1973, IEEE Trans. Inf. Theory.

[19]  Aaron D. Wyner,et al.  A theorem on the entropy of certain binary sequences and applications-II , 1973, IEEE Trans. Inf. Theory.

[20]  G. Chaitin A Theory of Program Size Formally Identical to Information Theory , 1975, JACM.

[21]  Thomas M. Cover,et al.  Compound Bayes Predictors for Sequences with Apparent Markov Structure , 1977, IEEE Transactions on Systems, Man, and Cybernetics.

[22]  S. Vardeman Admissible Solutions of Finite State Sequence Compound Decision Problems , 1978 .

[23]  Abraham Lempel,et al.  Compression of individual sequences via variable-rate coding , 1978, IEEE Trans. Inf. Theory.

[24]  Y. Nogami Thek-extended set-compound estimation problem in a nonregular family of distrubutions over [θ, θ+1) , 1979 .

[25]  P. Hall,et al.  Martingale Limit Theory and Its Application , 1980 .

[26]  S. Vardeman Admissible solutions of k-extended finite state set and sequence compound decision problems , 1980 .

[27]  Jorma Rissanen,et al.  Universal coding, information, prediction, and estimation , 1984, IEEE Trans. Inf. Theory.

[28]  H. Robbins Asymptotically Subminimax Solutions of Compound Statistical Decision Problems , 1985 .

[29]  Vladimir Vovk,et al.  Aggregating strategies , 1990, COLT '90.

[30]  R. Durrett Probability: Theory and Examples , 1993 .

[31]  Meir Feder,et al.  Gambling using a finite state machine , 1991, IEEE Trans. Inf. Theory.

[32]  Neri Merhav,et al.  Universal coding with minimum probability of codeword length overflow , 1991, IEEE Trans. Inf. Theory.

[33]  Neri Merhav,et al.  Universal prediction of individual sequences , 1992, IEEE Trans. Inf. Theory.

[34]  David Haussler,et al.  How to use expert advice , 1993, STOC.

[35]  N. Merhav,et al.  Universal Schemes for Sequential Decision from Individual Data Sequences , 1993, Proceedings. IEEE International Symposium on Information Theory.

[36]  Neri Merhav,et al.  Optimal sequential probability assignment for individual sequences , 1994, IEEE Trans. Inf. Theory.

[37]  E. Rio The Functional Law of the Iterated Logarithm for Stationary Strongly Mixing Sequences , 1995 .

[38]  Vladimir Vovk,et al.  A game of prediction with expert advice , 1995, COLT '95.

[39]  David Haussler,et al.  Sequential Prediction of Individual Sequences Under General Loss Functions , 1998, IEEE Trans. Inf. Theory.

[40]  Amir Dembo,et al.  Large Deviations Techniques and Applications , 1998 .

[41]  Neri Merhav,et al.  Universal Prediction , 1998, IEEE Trans. Inf. Theory.

[42]  G. Lugosi,et al.  On Prediction of Individual Sequences , 1998 .

[43]  Nicolò Cesa-Bianchi,et al.  On sequential prediction of individual sequences relative to a set of experts , 1998, COLT' 98.

[44]  Tsachy Weissman,et al.  On prediction of individual sequences relative to a set of experts in the presence of noise , 1999, COLT '99.

[45]  D. Haussler,et al.  Worst Case Prediction over Sequences under Log Loss , 1999 .

[46]  N. Merhav,et al.  Universal prediction of individual binary sequences in the presence of arbitrarily varying, memoryless additive noise , 2000, 2000 IEEE International Symposium on Information Theory (Cat. No.00CH37060).

[47]  Tsachy Weissman,et al.  Universal prediction of individual binary sequences in the presence of noise , 2001, IEEE Trans. Inf. Theory.