First- and Second-Order Coding Theorems for Mixed Memoryless Channels With General Mixture

This paper investigates the first- and second-order maximum achievable rates of codes with/without cost constraints for mixed channels whose channel law is characterized by a general mixture of (at most) uncountably many stationary and memoryless discrete channels. These channels are referred to as mixed memoryless channels with general mixture and include the class of mixed memoryless channels of finitely or countably memoryless channels as a special case. For the mixed memoryless channels with general mixture, the first-order coding theorem which gives a formula for the $\varepsilon $ -capacity is established, and then a direct part of the second-order coding theorem is provided. A subclass of mixed memoryless channels whose component channels can be ordered according to their capacity is introduced, and the first- and second-order coding theorems are established. It is shown that the established formulas reduce to several known formulas for restricted scenarios.

[1]  Richard E. Blahut,et al.  Computation of channel capacity and rate-distortion functions , 1972, IEEE Trans. Inf. Theory.

[2]  Karel Winkelbauer On the coding theorem for decomposable discrete information channels. II , 1971, Kybernetika.

[3]  Rudolf Ahlswede,et al.  The weak capacity of averaged channels , 1968 .

[4]  Sergio Verdú,et al.  A general formula for channel capacity , 1994, IEEE Trans. Inf. Theory.

[5]  H. Vincent Poor,et al.  Channel Coding Rate in the Finite Blocklength Regime , 2010, IEEE Transactions on Information Theory.

[6]  Masahito Hayashi,et al.  Information Spectrum Approach to Second-Order Coding Rate in Channel Coding , 2008, IEEE Transactions on Information Theory.

[7]  Masahito Hayashi,et al.  General formulas for capacity of classical-quantum channels , 2003, IEEE Transactions on Information Theory.

[8]  Giuseppe Durisi,et al.  Quasi-static SIMO fading channels at finite blocklength , 2013, 2013 IEEE International Symposium on Information Theory.

[9]  H. Vincent Poor,et al.  Dispersion of the Gilbert-Elliott Channel , 2011, IEEE Trans. Inf. Theory.

[10]  Hideki Yagi,et al.  Single-letter characterization of epsilon-capacity for mixed memoryless channels , 2014, 2014 IEEE International Symposium on Information Theory.

[11]  John C. Kieffer,et al.  $\epsilon$-Capacity of Binary Symmetric Averaged Channels , 2007, IEEE Transactions on Information Theory.

[12]  Te Sun Han,et al.  Second-Order Slepian-Wolf Coding Theorems for Non-Mixed and Mixed Sources , 2014, IEEE Trans. Inf. Theory.

[13]  Vincent Yan Fu Tan,et al.  A Tight Upper Bound for the Third-Order Asymptotics for Most Discrete Memoryless Channels , 2012, IEEE Transactions on Information Theory.

[14]  Vincent Yan Fu Tan,et al.  Second-Order Coding Rates for Channels With State , 2014, IEEE Transactions on Information Theory.

[15]  Te Sun Han,et al.  Second-Order Resolvability, Intrinsic Randomness, and Fixed-Length Source Coding for Mixed Sources: Information Spectrum Approach , 2011, IEEE Transactions on Information Theory.

[16]  Amiel Feinstein,et al.  A new basic theorem of information theory , 1954, Trans. IRE Prof. Group Inf. Theory.

[17]  Hideki Yagi,et al.  Channel dispersion for well-ordered mixed channels decomposed into memoryless channels , 2014, 2014 International Symposium on Information Theory and its Applications.

[18]  C. E. SHANNON,et al.  A mathematical theory of communication , 1948, MOCO.