On the capacity of Markov sources over noisy channels
暂无分享,去创建一个
[1] Paul H. Siegel,et al. Codes for Digital Recorders , 1998, IEEE Trans. Inf. Theory.
[2] David L. Neuhoff,et al. Coding for channels with cost constraints , 1996, IEEE Trans. Inf. Theory.
[3] Jack K. Wolf,et al. On runlength codes , 1988, IEEE Trans. Inf. Theory.
[4] Richard E. Blahut,et al. Computation of channel capacity and rate-distortion functions , 1972, IEEE Trans. Inf. Theory.
[5] John Cocke,et al. Optimal decoding of linear codes for minimizing symbol error rate (Corresp.) , 1974, IEEE Trans. Inf. Theory.
[6] C. E. SHANNON,et al. A mathematical theory of communication , 1948, MOCO.
[7] Hans-Andrea Loeliger,et al. On the information rate of binary-input channels with memory , 2001, ICC 2001. IEEE International Conference on Communications. Conference Record (Cat. No.01CH37240).
[8] Shlomo Shamai,et al. The intersymbol interference channel: lower bounds on capacity and channel precoding loss , 1996, IEEE Trans. Inf. Theory.
[9] Paul H. Siegel,et al. On the achievable information rates of finite state ISI channels , 2001, GLOBECOM'01. IEEE Global Telecommunications Conference (Cat. No.01CH37270).
[10] Shlomo Shamai,et al. Information rates for a discrete-time Gaussian channel with intersymbol interference and stationary inputs , 1991, IEEE Trans. Inf. Theory.
[11] Claude E. Shannon,et al. A Mathematical Theory of Communications , 1948 .
[12] Suguru Arimoto,et al. An algorithm for computing the capacity of arbitrary discrete memoryless channels , 1972, IEEE Trans. Inf. Theory.