Rate-distortion speech coding with a minimum discrimination information distortion measure

An information theory approach to the theory and practice of linear predictive coded (LPC) speech compression systems is developed. It is shown that a traditional LPC system can be viewed as a minimum distortion or nearest-neighbor system where the distortion measure is a minimum discrimination information between a speech process model and an observed frame of actual speech. This distortion measure is used in an algorithm for computer-aided design of block source codes subject to a fidelity criterion to obtain a 750-bits/s speech compression system that resembles an LPC system but has a much lower rate, a larger memory requirement, and requires no on-line LPC analysis. Quantitative and informal subjective comparisons are made among our system and LPC systems.

[1]  C. E. SHANNON,et al.  A mathematical theory of communication , 1948, MOCO.

[2]  Claude E. Shannon,et al.  The Mathematical Theory of Communication , 1950 .

[3]  M. Kupperman PROBABILITIES OF HYPOTHESES AND INFORMATION-STATISTICS IN SAMPLING FROM EXPONENTIAL-CLASS POPULATIONS , 1958 .

[4]  W. Rudin The extension problem for positive-definite functions , 1963 .

[5]  J. P. Burg,et al.  Maximum entropy spectral analysis. , 1967 .

[6]  B. Atal,et al.  Speech analysis and synthesis by linear prediction of the speech wave. , 1971, The Journal of the Acoustical Society of America.

[7]  Adriaan van den Bos,et al.  Alternative interpretation of maximum entropy spectral analysis (Corresp.) , 1971, IEEE Trans. Inf. Theory.

[8]  A. Gray,et al.  On autocorrelation equations as applied to speech analysis , 1973 .

[9]  John B. Anderson,et al.  Tree encoding of speech , 1975, IEEE Trans. Inf. Theory.

[10]  I. Csiszár $I$-Divergence Geometry of Probability Distributions and Minimization Problems , 1975 .

[11]  J. Makhoul,et al.  Quantization properties of transmission parameters in linear predictive systems , 1975 .

[12]  George S. Kang,et al.  600 bps Voice digitizer , 1976, ICASSP.

[13]  A. Gray,et al.  Distance measures for speech processing , 1976 .

[14]  T. Ulrych,et al.  Time series modeling and maximum entropy , 1976 .

[15]  R. Gray,et al.  Comparison of optimal quantizations of speech reflection coefficients , 1977 .

[16]  Robert M. Gray,et al.  Source Coding and Speech Compression , 1978 .

[17]  Robert M. Gray,et al.  A two-step speech compression system with vector quantizing , 1979, ICASSP.

[18]  W Gersch,et al.  Automatic classification of electroencephalograms: Kullback-Leibler nearest neighbor rules. , 1979, Science.

[19]  J. Reid Confidence limits and maximum entropy spectra , 1979 .

[20]  A. Gray,et al.  Implementation and comparison of two transformed reflection coefficient scalar quantization methods , 1980 .

[21]  R. Gray,et al.  Distortion measures for speech processing , 1980 .

[22]  Robert M. Gray,et al.  An Algorithm for Vector Quantizer Design , 1980, IEEE Trans. Commun..

[23]  Robert M. Gray,et al.  Locally Optimal Block Quantizer Design , 1980, Inf. Control..

[24]  Rodney W. Johnson,et al.  Axiomatic derivation of the principle of maximum entropy and the principle of minimum cross-entropy , 1980, IEEE Trans. Inf. Theory.

[25]  R. Gray,et al.  Speech coding based upon vector quantization , 1980, ICASSP.

[26]  J. Shore Minimum cross-entropy spectral analysis , 1981 .

[27]  R. Johnson,et al.  Properties of cross-entropy minimization , 1981, IEEE Trans. Inf. Theory.

[28]  Robert M. Gray,et al.  Minimum Cross-Entropy Pattern Classification and Cluster Analysis , 1982, IEEE Transactions on Pattern Analysis and Machine Intelligence.