Deep Learning for Decoding of Linear Codes - A Syndrome-Based Approach

We present a novel framework for applying deep neural networks (DNN) to soft decoding of linear codes at arbitrary block lengths. Unlike other approaches, our framework allows unconstrained DNN design, enabling the free application of powerful designs that were developed in other contexts. Our method is robust to overfitting that inhibits many competing methods, which follows from the exponentially large number of codewords required for their training. We achieve this by transforming the channel output before feeding it to the network, extracting only the syndrome of the hard decisions and the channel output reliabilities. We prove analytically that this approach does not involve any intrinsic performance penalty, and guarantees the generalization of performance obtained during training. Our best results are obtained using a recurrent neural network (RNN) architecture combined with simple preprocessing by permutation. We provide simulation results that demonstrate performance that sometimes approaches that of the ordered statistics decoding (OSD) algorithm.

[1]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[2]  O. Antoine,et al.  Theory of Error-correcting Codes , 2022 .

[3]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[4]  David Burshtein,et al.  Deep Learning Methods for Improved Decoding of Linear Codes , 2017, IEEE Journal of Selected Topics in Signal Processing.

[5]  Rüdiger L. Urbanke,et al.  Modern Coding Theory , 2008 .

[6]  Timothy J. O'Shea,et al.  An Introduction to Machine Learning Communications Systems , 2017, ArXiv.

[7]  F. MacWilliams,et al.  The Theory of Error-Correcting Codes , 1977 .

[8]  Igal Sason,et al.  Parity-Check Density Versus Performance of Binary Linear Block Codes: New Bounds and Applications , 2007, IEEE Transactions on Information Theory.

[9]  L. Litwin,et al.  Error control coding , 2001 .

[10]  Geoffrey E. Hinton,et al.  Speech recognition with deep recurrent neural networks , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.

[11]  Yair Be'ery,et al.  Near Maximum Likelihood Decoding with Deep Learning , 2018, ArXiv.

[12]  D. Burshtein,et al.  Upper bounds on the rate of LDPC codes , 2002, Proceedings IEEE International Symposium on Information Theory,.

[13]  L. G. Tallini,et al.  Neural nets for decoding error-correcting codes , 1995, IEEE Technical Applications Conference and Workshops. Northcon/95. Conference Record.

[14]  Jakov Snyders,et al.  Reliability-Based Syndrome Decoding of Linear Block Codes , 1998, IEEE Trans. Inf. Theory.

[15]  Jakob Hoydis,et al.  An Introduction to Deep Learning for the Physical Layer , 2017, IEEE Transactions on Cognitive Communications and Networking.

[16]  Yoshua Bengio,et al.  Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.

[17]  Yuen-Hsien Tseng,et al.  Neural Network Decoders for Linear Block Codes , 2002, Int. J. Comput. Eng. Sci..

[18]  Geoffrey E. Hinton,et al.  Deep Learning , 2015, Nature.

[19]  Jehoshua Bruck,et al.  Neural networks, error-correcting codes, and polynomials over the binary n -cube , 1989, IEEE Trans. Inf. Theory.

[20]  Rüdiger L. Urbanke,et al.  The capacity of low-density parity-check codes under message-passing decoding , 2001, IEEE Trans. Inf. Theory.

[21]  Shu Lin,et al.  Soft-decision decoding of linear block codes based on ordered statistics , 1994, IEEE Trans. Inf. Theory.

[22]  Yair Be'ery,et al.  Improved random redundant iterative HDPC decoding , 2009, IEEE Transactions on Communications.