Analysis of parallelization effects on textual data compression

This article deals with the problem of textual data compression speed up by dividing a bigger file into smaller ones which are simultaneously compressed. This procedure is called parallel computing and this experiment uses a processor with four cores. Parallel processing is applied on five different algorithms, two of these are entropy coders and others are dictionary coders. Algorithms are mutually compared by speed and performance depending on the number of cores used.

[1]  James A. Storer,et al.  A Parallel Architecture for High-Speed Data Compression , 1991, J. Parallel Distributed Comput..

[2]  Ian H. Witten,et al.  Lossless Compression for Text and Images , 1997 .

[3]  Ken Kennedy,et al.  Performance of parallel processors , 1989, Parallel Comput..

[4]  Nasir M. Rajpoot,et al.  Dictionary-Based Data Compression , 2003 .

[5]  Michal Zemlicka,et al.  Compression of small text files using syllables , 2006, Data Compression Conference (DCC'06).

[6]  Ida Pu Fundamental Data Compression , 2005 .

[7]  Abraham Lempel,et al.  A universal algorithm for sequential data compression , 1977, IEEE Trans. Inf. Theory.

[8]  James A. Storer,et al.  LZ-based image compression , 2001, Inf. Sci..

[9]  Mark R. Nelson,et al.  LZW data compression , 1989 .