A Study of Parallel Data Compression Using Proper Orthogonal Decomposition on the K Computer

The growing power of supercomputers continues to improve scientists' ability to model larger, more sophisticated problems in science with higher accuracy. An equally important ability is to make full use of the data output from the simulations to help clarify the modeled phenomena and facilitate the discovery of new phenomena. However, along with the scale of computation, the size of the resulting data has exploded; it becomes infeasible to output most of the data, which defeats the purpose of conducting large-scale simulations. In order to address this issue so that more data may be archived and studied, we have developed a scalable parallel data compression solution to reduce the size of large-scale data with low computational cost and minimal error. We use the proper orthogonal decomposition (POD) method to compress data because this method can effectively extract the main features from the data, and the resulting compressed data can be decompressed in linear time. Our implementation achieves high parallel efficiency with a binary load-distributed approach, which is similar to the binary-swap image composition method. This approach allows us to effectively use all of the processors and to reduce the interprocessor communication cost throughout the parallel compression calculations. The results of tests using the K computer indicate the superior performance of our design and implementation.

[1]  Enrico Magli,et al.  Low-complexity lossy compression of hyperspectral images via informed quantization , 2010, 2010 IEEE International Conference on Image Processing.

[2]  Kwan-Liu Ma,et al.  An Adaptive Prediction-Based Approach to Lossless Compression of Floating-Point Volume Data , 2012, IEEE Transactions on Visualization and Computer Graphics.

[3]  R. Venkata Rao,et al.  Advanced Modeling and Optimization of Manufacturing Processes: International Research and Development , 2013 .

[4]  Martin Isenburg,et al.  Fast and Efficient Compression of Floating-Point Data , 2006, IEEE Transactions on Visualization and Computer Graphics.

[5]  Jeremy Iverson,et al.  Fast and Effective Lossy Compression Algorithms for Scientific Datasets , 2012, Euro-Par.

[6]  Nikolay N. Ponomarenko,et al.  Lossy compression of images without visible distortions and its application , 2010, IEEE 10th INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING PROCEEDINGS.

[7]  Kwan-Liu Ma,et al.  Massively parallel volume rendering using 2-3 swap image compositing , 2008, HiPC 2008.

[8]  Radan Huth,et al.  The effect of various methodological options on the detection of leading modes of sea level pressure variability , 2006 .

[9]  Fenxiong Chen,et al.  Algorithm of Data Compression Based on Multiple Principal Component Analysis over the WSN , 2010, 2010 6th International Conference on Wireless Communications Networking and Mobile Computing (WiCOM).

[10]  Kwan-Liu Ma,et al.  Parallel volume rendering using binary-swap compositing , 1994, IEEE Computer Graphics and Applications.

[11]  Robert Latham,et al.  Compressing the Incompressible with ISABELA: In-situ Reduction of Spatio-temporal Data , 2011, Euro-Par.

[12]  Antonio J. Plaza,et al.  On the Impact of Lossy Compression on Hyperspectral Image Classification and Unmixing , 2011, IEEE Geoscience and Remote Sensing Letters.