Wireless Image Transmission Using Deep Source Channel Coding With Attention Modules

Recent research on joint source channel coding (JSCC) for wireless communications has achieved great success owing to the employment of deep learning (DL). However, the existing work on DL based JSCC usually trains the designed network in a specific signal-to-noise ratio (SNR) and applies the network for the scenario with the target SNR. A number of networks are required to cover the scenario with a broad range of SNRs, which is computational inefficiency (in the training stage) and requires large storage. These shortages hinder the use of DL based JSCC for real wireless scenarios. We propose a novel method called Attention DL based JSCC (ADJSCC) that can deal with different SNRs with a single neural network. This design is inspired by the resource assignment strategy in traditional JSCC, which dynamically adjusts the compression ratio in source coding and the channel coding rates according to the SNR. As a resource allocation scheme, Attention Mechanism allocates computing resources to more critical tasks, which naturally fits for the resource assignment strategy. Instead of applying the resource allocation strategy in traditional JSCC, the ADJSCC uses the channel-wise soft attention to scale features according to the information of SNR. We compare the ADJSCC method with the state-of-the-art DL based JSCC method through extensive experiments to demonstrate its adaptability, robustness and versatility. Compared with the existing methods, the proposed method takes less storage and is more robust in the burst channel.

[1]  Peter Vary,et al.  Transactions Papers source-optimized channel coding for digital transmission channels , 2005, IEEE Transactions on Communications.

[2]  Valero Laparra,et al.  End-to-end Optimized Image Compression , 2016, ICLR.

[3]  Dumitru Erhan,et al.  Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[4]  Biing-Hwang Juang,et al.  Channel Agnostic End-to-End Learning Based Communication Systems with Conditional GAN , 2018, 2018 IEEE Globecom Workshops (GC Wkshps).

[5]  Laurence B. Milstein,et al.  Joint Source-Channel Coding and Unequal Error Protection for Video Plus Depth , 2015, IEEE Signal Processing Letters.

[6]  Martín Abadi,et al.  TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems , 2016, ArXiv.

[7]  Khalid Sayood,et al.  Joint source/channel coding for variable length codes , 2000, IEEE Trans. Commun..

[8]  Stephan ten Brink,et al.  On deep learning-based channel decoding , 2017, 2017 51st Annual Conference on Information Sciences and Systems (CISS).

[9]  Xiaogang Wang,et al.  Residual Attention Network for Image Classification , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[10]  Stefano Ermon,et al.  Neural Joint Source-Channel Coding , 2018, ICML.

[11]  Deniz Gündüz,et al.  Successive Refinement of Images with Deep Joint Source-Channel Coding , 2019, 2019 IEEE 20th International Workshop on Signal Processing Advances in Wireless Communications (SPAWC).

[12]  Christopher D. Manning,et al.  Effective Approaches to Attention-based Neural Machine Translation , 2015, EMNLP.

[13]  Gang Sun,et al.  Squeeze-and-Excitation Networks , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[14]  C. E. SHANNON,et al.  A mathematical theory of communication , 1948, MOCO.

[15]  Alex Graves,et al.  Recurrent Models of Visual Attention , 2014, NIPS.

[16]  R. Gallager Information Theory and Reliable Communication , 1968 .

[17]  Himanshu Asnani,et al.  LEARN Codes: Inventing Low-Latency Codes via Recurrent Neural Networks , 2020, IEEE Journal on Selected Areas in Information Theory.

[18]  Jakob Hoydis,et al.  An Introduction to Deep Learning for the Physical Layer , 2017, IEEE Transactions on Cognitive Communications and Networking.

[19]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[20]  P. Viswanath,et al.  Turbo Autoencoder: Deep learning based channel codes for point-to-point communication channels , 2019, NeurIPS.

[21]  Bo Ai,et al.  Identification of Vehicle Obstruction Scenario Based on Machine Learning in Vehicle-to-vehicle Communications , 2020, 2020 IEEE 91st Vehicular Technology Conference (VTC2020-Spring).

[22]  Sumit Roy,et al.  Impact and mitigation of narrow-band radar interference in down-link LTE , 2015, 2015 IEEE International Conference on Communications (ICC).

[23]  C. Guillemot,et al.  Joint source-channel decoding of quasiarithmetic codes , 2004, Data Compression Conference, 2004. Proceedings. DCC 2004.

[24]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[25]  Valero Laparra,et al.  Density Modeling of Images using a Generalized Normalization Transformation , 2015, ICLR.

[26]  Jakob Hoydis,et al.  End-to-End Learning of Communications Systems Without a Channel Model , 2018, 2018 52nd Asilomar Conference on Signals, Systems, and Computers.

[27]  Deniz Gündüz,et al.  DeepJSCC-f: Deep Joint Source-Channel Coding of Images With Feedback , 2020, IEEE Journal on Selected Areas in Information Theory.

[28]  Yangmin Lee,et al.  Classification of node degree based on deep learning and routing method applied for virtual route assignment , 2017, Ad Hoc Networks.

[29]  Fady Alajaji,et al.  Joint Source–Channel Coding Error Exponent for Discrete Communication Systems With Markovian Memory , 2007, IEEE Transactions on Information Theory.

[30]  David Minnen,et al.  Variable Rate Image Compression with Recurrent Neural Networks , 2015, ICLR.

[31]  David Minnen,et al.  Channel-Wise Autoregressive Entropy Models for Learned Image Compression , 2020, 2020 IEEE International Conference on Image Processing (ICIP).

[32]  Yoshua Bengio,et al.  Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.

[33]  Michael S. Bernstein,et al.  ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.

[34]  Andriy Mnih,et al.  Variational Inference for Monte Carlo Objectives , 2016, ICML.

[35]  Jian Sun,et al.  Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[36]  In-So Kweon,et al.  CBAM: Convolutional Block Attention Module , 2018, ECCV.

[37]  Fady Alajaji,et al.  Error Exponents for Asymmetric Two-User Discrete Memoryless Source-Channel Systems , 2007, 2007 IEEE International Symposium on Information Theory.

[38]  Bernd Girod,et al.  Joint source channel coding of images with trellis coded quantization and convolutional codes , 1995, Proceedings., International Conference on Image Processing.

[39]  David Minnen,et al.  Joint Autoregressive and Hierarchical Priors for Learned Image Compression , 2018, NeurIPS.

[40]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[41]  Shi Jin,et al.  Deep Learning for Massive MIMO CSI Feedback , 2017, IEEE Wireless Communications Letters.

[42]  Bo Ai,et al.  Performance Evaluation of Autoencoder for Coding and Modulation in Wireless Communications , 2019, 2019 11th International Conference on Wireless Communications and Signal Processing (WCSP).