Asynchronous Slepian-Wolf code design

We consider asynchronous Slepian-Wolf coding where the two encoders may not have completely accurate timing information to synchronize their individual block code boundaries, and propose LDPC code design in this scenario. A new information-theoretic coding scheme based on source splitting is provided, which can achieve the entire asynchronous Slepian-Wolf rate region. Unlike existing methods based on source splitting, the proposed scheme does not require common randomness at the encoder and the decoder, or constructing super-letter from several individual symbols. Furthermore, we show that linear codes are sufficient for each coding step of this scheme. We subsequently design LDPC codes based on this new scheme, by applying the recently discovered source-channel code correspondence. Experimental results validate the effectiveness of the proposed method.

[1]  R. A. McDonald,et al.  Noiseless Coding of Correlated Information Sources , 1973 .

[2]  Imre Csiszár Linear codes for sources and source networks: Error exponents, universal coding , 1982, IEEE Trans. Inf. Theory.

[3]  Muriel Médard,et al.  Low-Complexity Approaches to Slepian–Wolf Near-Lossless Distributed Data Compression , 2006, IEEE Transactions on Information Theory.

[4]  Thomas M. Cover,et al.  A Proof of the Data Compression Theorem of Slepian and Wolf for Ergodic Sources , 1971 .

[5]  Patrick Mitran,et al.  Coding for the Slepian-Wolf problem with turbo codes , 2001, GLOBECOM'01. IEEE Global Telecommunications Conference (Cat. No.01CH37270).

[6]  Bernd Girod,et al.  Compression with side information using turbo codes , 2002, Proceedings DCC 2002. Data Compression Conference.

[7]  Zixiang Xiong,et al.  On code design for the Slepian-Wolf problem and lossless multiterminal networks , 2006, IEEE Transactions on Information Theory.

[8]  Ying Zhao,et al.  Compression of correlated binary sources using turbo codes , 2001, IEEE Communications Letters.

[9]  R. Urbanke,et al.  Asynchronous Slepian-Wolf coding via source-splitting , 1997, Proceedings of IEEE International Symposium on Information Theory.

[10]  I. Csiszár Information Theory , 1981 .

[11]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[12]  Toby Berger,et al.  Review of Information Theory: Coding Theorems for Discrete Memoryless Systems (Csiszár, I., and Körner, J.; 1981) , 1984, IEEE Trans. Inf. Theory.

[13]  Sae-Young Chung,et al.  On the design of low-density parity-check codes within 0.0045 dB of the Shannon limit , 2001, IEEE Communications Letters.

[14]  Rudolf Ahlswede,et al.  Source coding with side information and a converse for degraded broadcast channels , 1975, IEEE Trans. Inf. Theory.

[15]  Zixiang Xiong,et al.  Compression of binary sources with side information at the decoder using LDPC codes , 2002, IEEE Communications Letters.

[16]  Frans M. J. Willems Totally asynchronous Slepian-Wolf data compression , 1988, IEEE Trans. Inf. Theory.

[17]  Kannan Ramchandran,et al.  Distributed code constructions for the entire Slepian-Wolf rate region for arbitrarily correlated sources , 2004, Data Compression Conference, 2004. Proceedings. DCC 2004.

[18]  Jun Chen,et al.  On the Codebook-Level Duality Between Slepian-Woif Coding and Channel Coding , 2007, 2007 Information Theory and Applications Workshop.

[19]  Imre Csiszár,et al.  Information Theory - Coding Theorems for Discrete Memoryless Systems, Second Edition , 2011 .