Attention-based neural joint source-channel coding of text for point to point and broadcast channel

In this work, we consider the transmissions of structured data such as text over a noisy channel and correlated texts over a broadcast channel. As the separate source-channel coding principle no longer holds in such scenarios, we propose a joint source-channel coding scheme which is based on deep learning architecture. In order to enhance the convergence speed, we adopt the bidirectional gated recurrent unit at the encoder. For the decoder, to improve the recovery quality, we propose the following two types of strategies: (1) After a unidirectional neural network based decoder is used, a generative adversarial network is applied to train the whole joint source-channel coding framework and pointwise mutual information is added to the objective function of beam search process; (2) Rather than using a unidirectional neural network-based decoder, we develop a bidirectional neural network based and bidirectional attention mechanism integrated decoder to utilize past and future information. Experiments under different types of channels show that our schemes are superior to the existing deep learning joint source-channel coding method and in the case of low bit budget, long sentence length and small channel signal to noise ratio, our models are significantly superior to those of separate source-channel coding. In addition, we extend the proposed unidirectional and bidirectional decoders to the broadcast channel. Additionally, to improve the performance of unidirectional decoding, we utilize not only the correlation between adjacent words in the same text but also the correlation between words in different languages with the same meaning in the beam search process.

[1]  Christopher D. Manning,et al.  Effective Approaches to Attention-based Neural Machine Translation , 2015, EMNLP.

[2]  Slavche Pejoski,et al.  Joint source channel coding framework for real time H.264/AVC video transmission over wireless MIMO channels using outage probability , 2013, 2013 11th International Conference on Telecommunications in Modern Satellite, Cable and Broadcasting Services (TELSIKS).

[3]  Liang Xu,et al.  CLUECorpus2020: A Large-scale Chinese Corpus for Pre-training Language Model , 2020, ArXiv.

[4]  Xiang Zhang,et al.  Text Understanding from Scratch , 2015, ArXiv.

[5]  Ting Liu,et al.  Deep Learning-Based Belief Propagation Algorithm over Non-Binary Finite Fields , 2020, 2020 International Conference on Wireless Communications and Signal Processing (WCSP).

[6]  Jiajun Zhang,et al.  Synchronous Bidirectional Neural Machine Translation , 2019, TACL.

[7]  Bai Wang,et al.  Extracting Topics Based on Word2Vec and Improved Jaccard Similarity Coefficient , 2017, 2017 IEEE Second International Conference on Data Science in Cyberspace (DSC).

[8]  Thomas M. Cover,et al.  Broadcast channels , 1972, IEEE Trans. Inf. Theory.

[9]  Jing Zhang,et al.  Deep neural network-based underwater OFDM receiver , 2019, IET Commun..

[10]  Wei Chen,et al.  Improving Neural Machine Translation with Conditional Sequence Generative Adversarial Nets , 2017, NAACL.

[11]  Aman Jain,et al.  Information Retrieval using Cosine and Jaccard Similarity Measures in Vector Space Model , 2017 .

[12]  Alan Ritter,et al.  Adversarial Learning for Neural Dialogue Generation , 2017, EMNLP.

[13]  Zaid Omar,et al.  A review of hand gesture and sign language recognition techniques , 2017, International Journal of Machine Learning and Cybernetics.

[14]  Andrea J. Goldsmith,et al.  Variable Length Joint Source-Channel Coding of Text Using Deep Neural Networks , 2018, 2018 IEEE 19th International Workshop on Signal Processing Advances in Wireless Communications (SPAWC).

[15]  Stefano Ermon,et al.  Neural Joint Source-Channel Coding , 2018, ICML.

[16]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[17]  Jonathan Le Roux,et al.  Deep Unfolding: Model-Based Inspiration of Novel Deep Architectures , 2014, ArXiv.

[18]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[19]  Yoshua Bengio,et al.  Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.

[20]  Mohd Azri Mohd Izhar,et al.  Exploiting 2-Dimensional Source Correlation in Channel Decoding with Parameter Estimation , 2018 .

[21]  David A. Huffman,et al.  A method for the construction of minimum-redundancy codes , 1952, Proceedings of the IRE.

[22]  Andrea J. Goldsmith,et al.  Deep Learning for Joint Source-Channel Coding of Text , 2018, 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[23]  C. E. SHANNON,et al.  A mathematical theory of communication , 1948, MOCO.

[24]  Quoc V. Le,et al.  Sequence to Sequence Learning with Neural Networks , 2014, NIPS.

[25]  Shlomo Shamai,et al.  Superposition coding for side-information channels , 2006, IEEE Transactions on Information Theory.

[26]  I. Reed,et al.  Polynomial Codes Over Certain Finite Fields , 1960 .

[27]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[28]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[29]  Tie-Yan Liu,et al.  Adversarial Neural Machine Translation , 2017, ACML.

[30]  Ertem Tuncel,et al.  Wyner-Ziv Coding Over Broadcast Channels: Hybrid Digital/Analog Schemes , 2011, IEEE Transactions on Information Theory.

[31]  Jianfeng Gao,et al.  A Diversity-Promoting Objective Function for Neural Conversation Models , 2015, NAACL.

[32]  Henry D. Chadwick,et al.  Binary single-sideband phase-modulated communication systems (Corresp.) , 1972, IEEE Trans. Inf. Theory.

[33]  Abraham Lempel,et al.  A universal algorithm for sequential data compression , 1977, IEEE Trans. Inf. Theory.

[34]  Alex Graves,et al.  Sequence Transduction with Recurrent Neural Networks , 2012, ArXiv.

[35]  Jeya Sudharsan Shemona,et al.  Segmentation techniques for early cancer detection in red blood cells with deep learning-based classifier - a comparative approach , 2020, IET Image Process..

[36]  Luis Castedo,et al.  Experimental Evaluation of Analog Encoding for the Wireless Transmission of Still Images , 2019, 2019 IEEE 20th International Workshop on Signal Processing Advances in Wireless Communications (SPAWC).

[37]  Alex Graves,et al.  Supervised Sequence Labelling with Recurrent Neural Networks , 2012, Studies in Computational Intelligence.

[38]  Deniz Gündüz,et al.  Deep Joint Source-channel Coding for Wireless Image Transmission , 2018, ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[39]  Gonzalo Navarro,et al.  A guided tour to approximate string matching , 2001, CSUR.

[40]  Philipp Koehn,et al.  Two New Evaluation Datasets for Low-Resource Machine Translation: Nepali-English and Sinhala-English , 2019, ArXiv.

[41]  Simone A. Ludwig,et al.  Performance Evaluation of Deep Neural Networks Applied to Speech Recognition: RNN, LSTM and GRU , 2019, J. Artif. Intell. Soft Comput. Res..