Properties of the MMSE of “bad” codes
暂无分享,去创建一个
[1] Daniel Pérez Palomar,et al. Gradient of mutual information in linear vector Gaussian channels , 2006, IEEE Transactions on Information Theory.
[2] Shlomo Shamai,et al. MMSE of “Bad” Codes , 2013, IEEE Transactions on Information Theory.
[3] Hua Wang,et al. Gaussian Interference Channel Capacity to Within One Bit , 2007, IEEE Transactions on Information Theory.
[4] Te Sun Han,et al. A new achievable rate region for the interference channel , 1981, IEEE Trans. Inf. Theory.
[5] Shlomo Shamai,et al. Mutual information and minimum mean-square error in Gaussian channels , 2004, IEEE Transactions on Information Theory.
[6] S. Shamai,et al. Bounds on the MMSE of “bad” LDPC codes at rates above capacity , 2008, 2008 46th Annual Allerton Conference on Communication, Control, and Computing.
[7] A. Robert Calderbank,et al. In Praise of Bad Codes for Multi-Terminal Communications , 2010, ArXiv.
[8] Shlomo Shamai,et al. Information Theory On extrinsic information of good binary codes operating over Gaussian channels , 2007, Eur. Trans. Telecommun..
[9] Shlomo Shamai,et al. Performance Analysis of Linear Codes under Maximum-Likelihood Decoding: A Tutorial , 2006, Found. Trends Commun. Inf. Theory.
[10] Shlomo Shamai,et al. On MMSE properties and I-MMSE implications in parallel MIMO Gaussian channels , 2010, 2010 IEEE International Symposium on Information Theory.
[11] Shlomo Shamai,et al. Estimation in Gaussian Noise: Properties of the Minimum Mean-Square Error , 2010, IEEE Transactions on Information Theory.
[12] Thomas M. Cover,et al. Elements of Information Theory , 2005 .
[13] Shlomo Shamai,et al. Statistical Physics of Signal Estimation in Gaussian Noise: Theory and Examples of Phase Transitions , 2008, IEEE Transactions on Information Theory.