Neural network approaches to point lattice decoding

We characterize the complexity of the lattice decoding problem from a neural network perspective. The notion of Voronoi-reduced basis is introduced to restrict the space of solutions to a binary set. On the one hand, this problem is shown to be equivalent to computing a continuous piecewise linear (CPWL) function restricted to the fundamental parallelotope. On the other hand, it is known that any function computed by a ReLU feed-forward neural network is CPWL. As a result, we count the number of affine pieces in the CPWL decoding function to characterize the complexity of the decoding problem. It is exponential in the space dimension $n$, which induces shallow neural networks of exponential size. For structured lattices we show that folding, a technique equivalent to using a deep neural network, enables to reduce this complexity from exponential in $n$ to polynomial in $n$. Regarding unstructured MIMO lattices, in contrary to dense lattices many pieces in the CPWL decoding function can be neglected for quasi-optimal decoding on the Gaussian channel. This makes the decoding problem easier and it explains why shallow neural networks of reasonable size are more efficient with this category of lattices (in low to moderate dimensions).

[1]  Stephan ten Brink,et al.  On deep learning-based channel decoding , 2017, 2017 51st Annual Conference on Information Sciences and Systems (CISS).

[2]  Alexander Vardy,et al.  Closest point search in lattices , 2002, IEEE Trans. Inf. Theory.

[3]  Jakob Hoydis,et al.  An Introduction to Deep Learning for the Physical Layer , 2017, IEEE Transactions on Cognitive Communications and Networking.

[4]  Razvan Pascanu,et al.  On the number of response regions of deep feed forward networks with piece-wise linear activations , 2013, 1312.6098.

[5]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[6]  Ami Wiesel,et al.  Deep MIMO detection , 2017, 2017 IEEE 18th International Workshop on Signal Processing Advances in Wireless Communications (SPAWC).

[7]  Loïc Brunel,et al.  NEURAL LATTICE DECODERS , 2018, 2018 IEEE Global Conference on Signal and Information Processing (GlobalSIP).

[8]  Matus Telgarsky,et al.  Benefits of Depth in Neural Networks , 2016, COLT.

[9]  Razvan Pascanu,et al.  On the Number of Linear Regions of Deep Neural Networks , 2014, NIPS.

[10]  Raman Arora,et al.  Understanding Deep Neural Networks with Rectified Linear Units , 2016, Electron. Colloquium Comput. Complex..

[11]  Henri Cohen,et al.  A course in computational algebraic number theory , 1993, Graduate texts in mathematics.

[12]  Shafi Goldwasser,et al.  Complexity of lattice problems - a cryptographic perspective , 2002, The Kluwer international series in engineering and computer science.

[13]  Yair Be'ery,et al.  Learning to decode linear codes using deep learning , 2016, 2016 54th Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[14]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[15]  L. J. Boya,et al.  On Regular Polytopes , 2012, 1210.0601.

[16]  Henry Cohn,et al.  The sphere packing problem in dimension 8The sphere packing problem in dimension 8 , 2016, 1603.04246.

[17]  Neil J. A. Sloane,et al.  Low-dimensional lattices. VI. Voronoi reduction of three-dimensional lattices , 1992, Proceedings of the Royal Society of London. Series A: Mathematical and Physical Sciences.

[18]  Brendan McCane,et al.  Learning in deep architectures with folding transformations , 2013, The 2013 International Joint Conference on Neural Networks (IJCNN).

[19]  Loïc Brunel,et al.  Multilevel MIMO Detection with Deep Learning , 2018, 2018 52nd Asilomar Conference on Signals, Systems, and Computers.

[20]  W. Fischer,et al.  Sphere Packings, Lattices and Groups , 1990 .

[21]  Gregory Poltyrev,et al.  On coding without restrictions for the AWGN channel , 1993, IEEE Trans. Inf. Theory.