Binary Hypothesis Testing with Deterministic Finite-Memory Decision Rules

In this paper we consider the problem of binary hypothesis testing with finite memory systems. Let X1, X2, …. be a sequence of independent identically distributed Bernoulli random variables, with expectation p under ${{\mathcal{H}}_0}$ and q under ${{\mathcal{H}}_1}$. Consider a finite-memory deterministic machine with S states that updates its state Mn ∈ {1,2,…, S} at each time according to the rule Mn = f(Mn-1, Xn), where f is a deterministic time-invariant function. Assume that we let the process run for a very long time (n→∞), and then make our decision according to some mapping from the state space to the hypothesis space. The main contribution of this paper is a lower bound on the Bayes error probability Pe of any such machine. In particular, our findings show that the ratio between the maximal exponential decay rate of Pe with S for a deterministic machine and for a randomized one, can become unbounded, complementing a result by Hellman.

[1]  Martin J. Wainwright,et al.  Information-theoretic lower bounds for distributed statistical estimation with communication constraints , 2013, NIPS.

[2]  Himanshu Tyagi,et al.  Distributed Simulation and Distributed Inference , 2018, Electron. Colloquium Comput. Complex..

[3]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[4]  David P. Woodruff,et al.  Communication lower bounds for statistical estimation problems via a distributed data processing inequality , 2015, STOC.

[5]  Gregory Valiant,et al.  Memory, Communication, and Statistical Queries , 2016, COLT.

[6]  Ohad Shamir,et al.  Detecting Correlations with Little Memory and Communication , 2018, COLT 2018.

[7]  John C. Duchi,et al.  Minimax rates for memory-bounded sparse linear regression , 2015, COLT.

[8]  T. Cover,et al.  On Memory Saved by Randomization , 1971 .

[9]  Ohad Shamir,et al.  Space lower bounds for linear prediction in the streaming model , 2019, COLT.

[10]  Maxim Raginsky,et al.  Information-Theoretic Lower Bounds on Bayes Risk in Decentralized Estimation , 2016, IEEE Transactions on Information Theory.

[11]  B. Shubert,et al.  Testing a simple symmetric hypothesis by a finite-memory deterministic algorithm , 1973, IEEE Trans. Inf. Theory.

[12]  H Robbins,et al.  A SEQUENTIAL DECISION PROBLEM WITH A FINITE MEMORY. , 1956, Proceedings of the National Academy of Sciences of the United States of America.

[13]  Yanjun Han,et al.  A Geometric Characterization of Fisher Information from Quantized Samples with Applications to Distributed Statistical Estimation , 2018, 2018 56th Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[14]  Vatsal Sharan,et al.  Memory-sample tradeoffs for linear regression with small error , 2019, STOC.

[15]  John N. Tsitsiklis,et al.  On Learning With Finite Memory , 2012, IEEE Transactions on Information Theory.

[16]  Thomas M. Cover,et al.  A Note on the Two-Armed Bandit Problem with Finite Memory , 1968, Inf. Control..

[17]  Yanjun Han,et al.  Geometric Lower Bounds for Distributed Parameter Estimation Under Communication Constraints , 2018, IEEE Transactions on Information Theory.

[18]  Ran Raz,et al.  Fast Learning Requires Good Memory: A Time-Space Lower Bound for Parity Learning , 2016, 2016 IEEE 57th Annual Symposium on Foundations of Computer Science (FOCS).

[19]  B. Chandrasekaran,et al.  Finite-memory hypothesis testing-A critique (Corresp.) , 1970, IEEE Trans. Inf. Theory.

[20]  Martin E. Hellman,et al.  The effects of randomization on finite-memory decision schemes , 1972, IEEE Trans. Inf. Theory.

[21]  T. Cover Hypothesis Testing with Finite Statistics , 1969 .

[22]  William Feller,et al.  An Introduction to Probability Theory and Its Applications , 1967 .

[23]  Feller William,et al.  An Introduction To Probability Theory And Its Applications , 1950 .

[24]  William Feller,et al.  An Introduction to Probability Theory and Its Applications, Vol. 2 , 1967 .