Interval and recency rank source coding: Two on-line adaptive variable-length schemes

In the schemes presented the encoder maps each message into a codeword in a prefix-free codeword set. In interval encoding the codeword is indexed by the interval since the last previous occurrence of that message, and the codeword set must be countably infinite. In recency rank encoding the codeword is indexed by the number of distinct messages in that interval, and there must be no fewer codewords than messages. The decoder decodes each codeword on receipt. Users need not know message probabilities, but must agree on indexings, of the codeword set in an order of increasing length and of the message set in some arbitrary order. The average codeword length over a communications bout is never much larger than the value for an off-line scheme which maps the j th most frequent message in the bout into the j th shortest codeword in the given set, and is never too much larger than the value for off-line Huffman encoding of messages into the best codeword set for the bout message frequencies. Both schemes can do much better than Huffman coding when successive selections of each message type cluster much more than in the independent case.

[1]  Michael Rodeh,et al.  Economical encoding of commas between strings , 1978, CACM.

[2]  Peter Elias,et al.  Universal codeword sets and representations of the integers , 1975, IEEE Trans. Inf. Theory.

[3]  Jeffrey Scott Vitter,et al.  Design and analysis of dynamic Huffman coding , 1985, 26th Annual Symposium on Foundations of Computer Science (sfcs 1985).

[4]  James A. Storer,et al.  Parallel algorithms for data compression , 1985, JACM.

[5]  Robert E. Tarjan,et al.  A Locally Adaptive Data , 1986 .

[6]  Aaron D. Wyner,et al.  An Upper Bound on the Entropy Series , 1972, Inf. Control..

[7]  Donald E. Knuth,et al.  The art of computer programming: sorting and searching (volume 3) , 1973 .

[8]  Jacob Ziv,et al.  Coding theorems for individual sequences , 1978, IEEE Trans. Inf. Theory.

[9]  Michael Rodeh,et al.  Linear Algorithm for Data Compression via String Matching , 1981, JACM.

[10]  Terry A. Welch,et al.  A Technique for High-Performance Data Compression , 1984, Computer.

[11]  Peter Elias Minimax optimal universal codeword sets , 1983, IEEE Trans. Inf. Theory.

[12]  Andrew Chi-Chih Yao,et al.  An Almost Optimal Algorithm for Unbounded Searching , 1976, Inf. Process. Lett..

[13]  Alberto Leon-Garcia,et al.  A source matching approach to finding minimax codes , 1980, IEEE Trans. Inf. Theory.

[14]  Abraham Lempel,et al.  Compression of individual sequences via variable-rate coding , 1978, IEEE Trans. Inf. Theory.

[15]  Abraham Lempel,et al.  A universal algorithm for sequential data compression , 1977, IEEE Trans. Inf. Theory.

[16]  Robert E. Tarjan,et al.  Amortized efficiency of list update and paging rules , 1985, CACM.

[17]  Donald E. Knuth,et al.  Dynamic Huffman Coding , 1985, J. Algorithms.

[18]  Robert G. Gallager,et al.  Variations on a theme by Huffman , 1978, IEEE Trans. Inf. Theory.

[19]  Peter Elias Minimum Times and Memories Needed to Compute the Values of a Function , 1974, J. Comput. Syst. Sci..

[20]  Abraham Lempel,et al.  On the Complexity of Finite Sequences , 1976, IEEE Trans. Inf. Theory.

[21]  Jorma Rissanen,et al.  Minimax codes for finite alphabets (Corresp.) , 1978, IEEE Trans. Inf. Theory.