The minimax distortion redundancy in noisy source coding

Consider the problem of finite-rate filtering of a discrete memoryless process {X/sub i/}/sub i/spl ges/1/ based on its noisy observation sequence {Z/sub i/}/sub i/spl ges/1/, which is the output of a discrete memoryless channel (DMC) whose input is {X/sub i/}/sub i/spl ges/1/. When the distribution of the pairs (X/sub i/,Z/sub i/), P/sub X,Z/, is known, and for a given distortion measure, the solution to this problem is well known to be given by classical rate-distortion theory upon the introduction of a modified distortion measure. We address the case where P/sub X,Z/, rather than being completely specified, is only known to belong to some set /spl Lambda/. For a fixed encoding rate R, we look at the worst case, over all /spl theta//spl isin//spl Lambda/, of the difference between the expected distortion of a given scheme which is not allowed to depend on the active source /spl theta//spl isin//spl Lambda/ and the value of the distortion-rate function at R corresponding to the noisy source /spl theta/. We study the minimum attainable value achievable by any scheme operating at rate R for this worst case quantity, denoted by D(/spl Lambda/, R). Linking this problem and that of source coding under several distortion measures, we prove a coding theorem for the latter problem and apply it to characterize D(/spl Lambda/, R) for the case where all members of /spl Lambda/ share the same noisy marginal. For the case of a general /spl Lambda/, we obtain a single-letter characterization of D(/spl Lambda/, R) for the finite-alphabet case. This gives, in particular, a necessary and sufficient condition on the set /spl Lambda/ for the existence of a coding scheme which is universally optimal for all members of /spl Lambda/ and characterizes the approximation-estimation tradeoff for statistical modeling of noisy source coding problems. Finally, we obtain D(/spl Lambda/, R) in closed form for cases where /spl Lambda/ consists of distributions on the (channel) input-output pair of a Bernoulli source corrupted by a binary-symmetric channel (BSC). In particular, for the case where /spl Lambda/ consists of two sources: the all-zero source corrupted by a BSC with crossover probability r and the Bernoulli(r) source with a noise-free channel; we find that universality becomes increasingly hard with increasing rate.

[1]  Elza Erkip,et al.  The Efficiency of Investment Information , 1998, IEEE Trans. Inf. Theory.

[2]  Toby Berger,et al.  Rate distortion theory : a mathematical basis for data compression , 1971 .

[3]  Jacob Ziv,et al.  Distortion-rate theory for individual sequences , 1980, IEEE Trans. Inf. Theory.

[4]  Jorma Rissanen,et al.  Universal coding, information, prediction, and estimation , 1984, IEEE Trans. Inf. Theory.

[5]  Andrew R. Barron,et al.  Minimax redundancy for the class of memoryless sources , 1997, IEEE Trans. Inf. Theory.

[6]  Tsachy Weissman,et al.  Tradeoffs between the excess-code-length exponent and the excess-distortion exponent in lossy source coding , 2002, IEEE Trans. Inf. Theory.

[7]  R. G. Gallager,et al.  Coding of Sources With Unknown Statistics- Part II: Distortion Relative to a Fidelity Criterion , 1972 .

[8]  A. Dembo,et al.  A Topological Criterion for Hypothesis Testing , 1994 .

[9]  Daniel J. Costello,et al.  Universal lossless coding for sources with repeating statistics , 2004, IEEE Transactions on Information Theory.

[10]  D. Pollard Convergence of stochastic processes , 1984 .

[11]  R. Gallager Information Theory and Reliable Communication , 1968 .

[12]  Tamás Linder,et al.  Rates of convergence in the source coding theorem, in empirical quantizer design, and in universal lossy source coding , 1994, IEEE Trans. Inf. Theory.

[13]  David L. Neuhoff,et al.  Fixed rate universal block source coding with a fidelity criterion , 1975, IEEE Trans. Inf. Theory.

[14]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[15]  Jacob Ziv,et al.  Coding of sources with unknown statistics-II: Distortion relative to a fidelity criterion , 1972, IEEE Trans. Inf. Theory.

[16]  Jack K. Wolf,et al.  Transmission of noisy information to a noisy receiver with minimum distortion , 1970, IEEE Trans. Inf. Theory.

[17]  Jan-Erik Stjernvall Dominance - A relation between distortion measures , 1983, IEEE Trans. Inf. Theory.

[18]  Tamás Linder,et al.  The minimax distortion redundancy in empirical quantizer design , 1997, Proceedings of IEEE International Symposium on Information Theory.

[19]  Amir Dembo,et al.  Large Deviations Techniques and Applications , 1998 .

[20]  Balas K. Natarajan Filtering random noise via data compression , 1993, [Proceedings] DCC `93: Data Compression Conference.

[21]  Francis Comets,et al.  Large deviation estimates for a conditional probability distribution. Applications to random interaction Gibbs measures , 1989 .

[22]  R. Dudley,et al.  Uniform Central Limit Theorems: Notation Index , 2014 .

[23]  A. Lapidoth On the role of mismatch in rate distortion theory , 1995, Proceedings of 1995 IEEE International Symposium on Information Theory.

[24]  Michelle Effros,et al.  A vector quantization approach to universal noiseless coding and quantization , 1996, IEEE Trans. Inf. Theory.

[25]  Tsachy Weissman,et al.  On limited-delay lossy coding and filtering of individual sequences , 2002, IEEE Trans. Inf. Theory.

[26]  En-Hui Yang,et al.  Simple universal lossy data compression schemes derived from the Lempel-Ziv algorithm , 1996, IEEE Trans. Inf. Theory.

[27]  Hans S. Witsenhausen,et al.  Indirect rate distortion problems , 1980, IEEE Trans. Inf. Theory.

[28]  Robert M. Gray,et al.  A unified approach for encoding clean and noisy sources by means of waveform and autoregressive model vector quantization , 1988, IEEE Trans. Inf. Theory.

[29]  Lee D. Davisson,et al.  Universal noiseless coding , 1973, IEEE Trans. Inf. Theory.

[30]  Abbas El Gamal,et al.  Achievable rates for multiple descriptions , 1982, IEEE Trans. Inf. Theory.