Second-Order Asymptotics of Universal JSCC for Arbitrary Sources and Additive Channels

We consider a universal joint source channel coding (JSCC) scheme to transmit an arbitrary memoryless source over an arbitrary additive channel. We adopt an architecture that consists of Gaussian codebooks for both the source reproduction sequences and channel codewords. The natural minimum Euclidean distance encoder and decoder, however, need to be judiciously modified to ensure universality as well as to obtain the best (highest) possible communication rates. In particular, we consider the analogue of an unequal error (or message) protection scheme in which all sources are partitioned into disjoint power type classes. We also regularize the nearest neighbor decoder so an appropriate measure of the size of each power type class is taken into account in the decoding strategy. For such an architecture, we derive ensemble tight second-order asymptotics.

[1]  Vincent Y. F. Tan,et al.  The Dispersion of Nearest-Neighbor Decoding for Additive Non-Gaussian Channels , 2017, IEEE Trans. Inf. Theory.

[2]  Sergio Verdú,et al.  Lossy Joint Source-Channel Coding in the Finite Blocklength Regime , 2012, IEEE Transactions on Information Theory.

[3]  Mehul Motani,et al.  The Dispersion of Universal Joint Source-Channel Coding for Arbitrary Sources and Additive Channels , 2017, arXiv.org.

[4]  Oliver Kosut,et al.  Third-order coding rate for universal compression of Markov sources , 2015, 2015 IEEE International Symposium on Information Theory (ISIT).

[5]  Mehul Motani,et al.  The Dispersion of Mismatched Joint Source-Channel Coding for Arbitrary Sources and Additive Channels , 2017, IEEE Transactions on Information Theory.

[6]  Mehul Motani,et al.  Refined Asymptotics for Rate-Distortion Using Gaussian Codebooks for Arbitrary Sources , 2019, IEEE Transactions on Information Theory.

[7]  H. Vincent Poor,et al.  Channel Coding Rate in the Finite Blocklength Regime , 2010, IEEE Transactions on Information Theory.

[8]  Cheng Chang,et al.  Joint source-channel with side information coding error exponents , 2009, 2009 47th Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[9]  Yuval Kochman,et al.  The dispersion of joint source-channel coding , 2011, 2011 49th Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[10]  A. Lapidoth On the role of mismatch in rate distortion theory , 1995, Proceedings of 1995 IEEE International Symposium on Information Theory.

[11]  Robert G. Gallager,et al.  The random coding bound is tight for the average code (Corresp.) , 1973, IEEE Trans. Inf. Theory.

[12]  Amos Lapidoth,et al.  Nearest neighbor decoding for additive non-Gaussian noise channels , 1996, IEEE Trans. Inf. Theory.

[13]  Mehul Motani,et al.  Second-Order and Moderate Deviations Asymptotics for Successive Refinement , 2016, IEEE Transactions on Information Theory.

[14]  Aaron D. Wyner,et al.  Coding Theorems for a Discrete Source With a Fidelity CriterionInstitute of Radio Engineers, International Convention Record, vol. 7, 1959. , 1993 .

[15]  A. Lapidoth On the role of mismatch in rate distortion theory , 1997, IEEE Trans. Inf. Theory.