“Real” Slepian-Wolf codes

We provide a novel achievability proof of the Slepian-Wolf theorem for i.i.d. sources over finite alphabets. We demonstrate that random codes that are linear over the real field achieve the classical Slepian-Wolf rate region. For finite alphabets we show that decoding is equivalent to solving an integer program. The techniques used may be of independent interest for code design for a wide class of information theory problems, and for the field of compressed sensing.

[1]  Ying Zhao,et al.  Compression of correlated binary sources using turbo codes , 2001, IEEE Communications Letters.

[2]  Frank E. Grubbs,et al.  An Introduction to Probability Theory and Its Applications , 1951 .

[3]  Emmanuel J. Candès,et al.  Robust uncertainty principles: exact signal reconstruction from highly incomplete frequency information , 2004, IEEE Transactions on Information Theory.

[4]  Imre Csiszár Linear codes for sources and source networks: Error exponents, universal coding , 1982, IEEE Trans. Inf. Theory.

[5]  Kannan Ramchandran,et al.  PRISM: A new robust video coding architecture based on distributed compression principles , 2002 .

[6]  Muriel Médard,et al.  On some new approaches to practical Slepian-Wolf compression inspired by channel coding , 2004, Data Compression Conference, 2004. Proceedings. DCC 2004.

[7]  Rüdiger L. Urbanke,et al.  Lattice Codes Can Achieve Capacity on the AWGN Channel , 1998, IEEE Trans. Inf. Theory.

[8]  P. A. P. Moran,et al.  An introduction to probability theory , 1968 .

[9]  Kannan Ramchandran,et al.  Distributed compression in a dense microsensor network , 2002, IEEE Signal Process. Mag..

[10]  Imre Csiszár,et al.  Common randomness and secret key generation with a helper , 2000, IEEE Trans. Inf. Theory.

[11]  A. J. Hoffman The Role of Unimodularity in Applying Linear Inequalities to Combinatorial Theorems , 1979 .

[12]  P. Narayan,et al.  Common randomness and secret key generation with a helper , 1997, Proceedings of IEEE International Symposium on Information Theory.

[13]  Imre Csiszár,et al.  Towards a general theory of source networks , 1980, IEEE Trans. Inf. Theory.

[14]  William Feller,et al.  An Introduction to Probability Theory and Its Applications , 1967 .

[15]  Emmanuel J. Candès,et al.  Decoding by linear programming , 2005, IEEE Transactions on Information Theory.

[16]  Robert B. Ash,et al.  Information Theory , 2020, The SAGE International Encyclopedia of Mass Media and Society.

[17]  Jack K. Wolf,et al.  Noiseless coding of correlated information sources , 1973, IEEE Trans. Inf. Theory.

[18]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[19]  Sang Joon Kim,et al.  A Mathematical Theory of Communication , 2006 .