A beta-beta achievability bound with applications

A channel coding achievability bound expressed in terms of the ratio between two Neyman-Pearson β functions is proposed. This bound is the dual of a converse bound established earlier by Polyanskiy and Verdú (2014). The new bound turns out to simplify considerably the analysis in situations where the channel output distribution is not a product distribution, for example due to a cost constraint or a structural constraint (such as orthogonality or constant composition) on the channel inputs. Connections to existing bounds in the literature are discussed. The bound is then used to derive 1) the channel dispersion of additive non-Gaussian noise channels with random Gaussian codebooks, 2) the channel dispersion of an exponential-noise channel, 3) a second-order expansion for the minimum energy per bit of an additive white Gaussian noise channel, and 4) a lower bound on the maximum coding rate of a multiple-input multiple-output Rayleigh-fading channel with perfect channel state information at the receiver, which is the tightest known achievability result.

[1]  Yury Polyanskiy,et al.  Saddle Point in the Minimax Converse for Channel Coding , 2013, IEEE Transactions on Information Theory.

[2]  Sergio Verdú,et al.  Spectral efficiency in the wideband regime , 2002, IEEE Trans. Inf. Theory.

[3]  Sergio Verdú,et al.  Bits through queues , 1994, Proceedings of 1994 IEEE International Symposium on Information Theory.

[4]  Peter Harremoës,et al.  Rényi Divergence and Kullback-Leibler Divergence , 2012, IEEE Transactions on Information Theory.

[5]  Sergio Verdú,et al.  Empirical Distribution of Good Channel Codes With Nonvanishing Error Probability , 2014, IEEE Transactions on Information Theory.

[6]  Albert Guillén i Fàbregas,et al.  Mismatched Decoding: Error Exponents, Second-Order Rates and Saddlepoint Approximations , 2013, IEEE Transactions on Information Theory.

[7]  Claude E. Shannon,et al.  Certain Results in Coding Theory for Noisy Channels , 1957, Inf. Control..

[8]  William Feller,et al.  An Introduction to Probability Theory and Its Applications , 1967 .

[9]  Oliver Kosut,et al.  Third-order coding rate for universal compression of Markov sources , 2015, 2015 IEEE International Symposium on Information Theory (ISIT).

[10]  Richard E. Blahut,et al.  Computation of channel capacity and rate-distortion functions , 1972, IEEE Trans. Inf. Theory.

[11]  Sergio Verdu,et al.  Multiuser Detection , 1998 .

[12]  Suguru Arimoto,et al.  An algorithm for computing the capacity of arbitrary discrete memoryless channels , 1972, IEEE Trans. Inf. Theory.

[13]  Amos Lapidoth,et al.  Capacity bounds via duality with applications to multiple-antenna systems on flat-fading channels , 2003, IEEE Trans. Inf. Theory.

[14]  Andrew C. Singer,et al.  Finite block-length achievable rates for queuing timing channels , 2011, 2011 IEEE Information Theory Workshop.

[15]  Sergio Verdú,et al.  On channel capacity per unit cost , 1990, IEEE Trans. Inf. Theory.

[16]  Amiel Feinstein,et al.  A new basic theorem of information theory , 1954, Trans. IRE Prof. Group Inf. Theory.

[17]  J. Nicholas Laneman,et al.  A Second-Order Achievable Rate Region for Gaussian Multi-Access Channels via a Central Limit Theorem for Functions , 2015, IEEE Transactions on Information Theory.

[18]  Roger Colbeck,et al.  Simple channel coding bounds , 2009, 2009 IEEE International Symposium on Information Theory.

[19]  H. Vincent Poor,et al.  Channel Coding Rate in the Finite Blocklength Regime , 2010, IEEE Transactions on Information Theory.

[20]  D. A. Bell,et al.  Information Theory and Reliable Communication , 1969 .

[21]  E. S. Pearson,et al.  On the Problem of the Most Efficient Tests of Statistical Hypotheses , 1933 .

[22]  S. Verdú,et al.  Arimoto channel coding converse and Rényi divergence , 2010, 2010 48th Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[23]  Tadeusz Inglot,et al.  Asymptotic optimality of new adaptive test in regression model , 2006 .

[24]  Sergio Verdú,et al.  Scalar coherent fading channel: Dispersion analysis , 2011, 2011 IEEE International Symposium on Information Theory Proceedings.

[25]  Shlomo Shamai,et al.  The empirical distribution of good codes , 1997, IEEE Trans. Inf. Theory.

[26]  Imre Csiszár,et al.  Information Theory - Coding Theorems for Discrete Memoryless Systems, Second Edition , 2011 .

[27]  E. S. Pearson,et al.  On the Problem of the Most Efficient Tests of Statistical Hypotheses , 1933 .

[28]  H. Vincent Poor,et al.  Minimum Energy to Send $k$ Bits Through the Gaussian Channel With and Without Feedback , 2011, IEEE Transactions on Information Theory.

[29]  Yihong Wu,et al.  Wasserstein Continuity of Entropy and Outer Bounds for Interference Channels , 2015, IEEE Transactions on Information Theory.

[30]  Sergio Verdu,et al.  The exponential distribution in information theory , 1996 .

[31]  Sergio Verdú,et al.  Functional Properties of Minimum Mean-Square Error and Mutual Information , 2012, IEEE Transactions on Information Theory.

[32]  P. Massart,et al.  Adaptive estimation of a quadratic functional by model selection , 2000 .