Generalized Random Gilbert-Varshamov Codes

We introduce a random coding technique for transmission over discrete memoryless channels, reminiscent of the basic construction attaining the Gilbert–Varshamov bound for codes in Hamming spaces. The code construction is based on drawing codewords recursively from a fixed type class, in such a way that a newly generated codeword must be at a certain minimum distance from all previously chosen codewords, according to some generic distance function. We derive an achievable error exponent for this construction and prove its tightness with respect to the ensemble average. We show that the exponent recovers the Csiszár and Körner exponent as a special case, which is known to be at least as high as both the random-coding and expurgated exponents, and we establish the optimality of certain choices of the distance function. In addition, for additive distances and decoding metrics, we present an equivalent dual expression, along with a generalization to infinite alphabets via cost-constrained random coding.

[1]  Dominique de Caen,et al.  A lower bound on the probability of a union , 1997, Discret. Math..

[2]  Li Peng,et al.  Expurgated Random-Coding Ensembles: Exponents, Refinements, and Connections , 2013, IEEE Transactions on Information Theory.

[3]  Albert Guillén i Fàbregas,et al.  Mismatched Decoding: Error Exponents, Second-Order Rates and Saddlepoint Approximations , 2013, IEEE Transactions on Information Theory.

[4]  H. Vincent Poor,et al.  Channel Coding Rate in the Finite Blocklength Regime , 2010, IEEE Transactions on Information Theory.

[5]  Imre Csiszár,et al.  Graph decomposition: A new key to coding theorems , 1981, IEEE Trans. Inf. Theory.

[6]  R. Gallager Information Theory and Reliable Communication , 1968 .

[7]  Ari Trachenberg Error-Correcting Codes on Graphs: Lexicodes, Trellises and Factor Graphs , 1999 .

[8]  R. Brualdi,et al.  Greedy Codes , 1993, Proceedings. IEEE International Symposium on Information Theory.

[9]  G. Kaplan,et al.  On Information Rates for Mismatched Decoders , 1993, Proceedings. IEEE International Symposium on Information Theory.

[10]  Alexander Barg,et al.  Random codes: Minimum distances and error exponents , 2002, IEEE Trans. Inf. Theory.

[11]  Neri Merhav The Generalized Stochastic Likelihood Decoder: Random Coding and Expurgated Bounds , 2017, IEEE Trans. Inf. Theory.

[12]  Emre Telatar,et al.  Mismatched decoding revisited: General alphabets, channels with memory, and the wide-band limit , 2000, IEEE Trans. Inf. Theory.

[13]  Neri Merhav,et al.  Exact Random Coding Exponents for Erasure Decoding , 2011, IEEE Transactions on Information Theory.

[14]  Imre Csiszár,et al.  Channel capacity for a given decoding metric , 1995, IEEE Trans. Inf. Theory.

[15]  Richard E. Blahut,et al.  Composition bounds for channel block codes , 1977, IEEE Trans. Inf. Theory.

[16]  N. J. A. Sloane,et al.  Lexicographic codes: Error-correcting codes from game theory , 1986, IEEE Trans. Inf. Theory.

[17]  Imre Csiszár,et al.  Information Theory - Coding Theorems for Discrete Memoryless Systems, Second Edition , 2011 .

[18]  V. I. Siforov On noise stability of a system with error-correcting codes , 1956, IRE Trans. Inf. Theory.

[19]  Anelia Somekh-Baruch On Achievable Rates and Error Exponents for Channels With Mismatched Decoding , 2015, IEEE Transactions on Information Theory.

[20]  Joseph Y. N. Hui,et al.  Fundamental issues of multiple accessing , 1983 .

[21]  Albert Guillén i Fàbregas,et al.  Ensemble-tight error exponents for mismatched decoders , 2012, 2012 50th Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[22]  Neri Merhav,et al.  List Decoding—Random Coding Exponents and Expurgated Exponents , 2013, IEEE Transactions on Information Theory.

[23]  E. Gilbert A comparison of signalling alphabets , 1952 .