Collective I/O Performance on the Santos Dumont Supercomputer

The historical gap between processing and data access speeds causes many applications to spend a large portion of their execution on I/O operations. From the point of view of a large-scale, expensive, supercomputer, it is important to ensure applications achieve the best I/O performance to promote an efficient usage of the machine. In this paper, we evaluate the I/O infrastructure of the Santos Dumont supercomputer, the largest one from Latin America. More specifically, we investigate the performance of collective I/O operations. By conducting an analysis of a scientific application that uses the machine, we identify large performance differences between the available MPI implementations. We then further study the observed phenomenon using the BT-IO and IOR benchmarks, in addition to a custom microbenchmark. We conclude that the customized MPI implementation by Bull (used by more than 20% of the jobs) presents the worst performance for small collective write operations. Our results are being used to help the Santos Dumont users to achieve the best performance for their applications. Additionally, by investigating the observed phenomenon, we provide information to help improve future MPI-IO collective write implementations.

[1]  Scott Klasky,et al.  Characterizing output bottlenecks in a supercomputer , 2012, 2012 International Conference for High Performance Computing, Networking, Storage and Analysis.

[2]  Robert Latham,et al.  High performance file I/O for the Blue Gene/L supercomputer , 2006, The Twelfth International Symposium on High-Performance Computer Architecture, 2006..

[3]  Jean Luca Bez,et al.  A Checkpoint of Research on Parallel I/O for High-Performance Computing , 2018, ACM Comput. Surv..

[4]  Robert Latham,et al.  Understanding and improving computational science storage access through continuous characterization , 2011, MSST.

[5]  Marianne Winslett,et al.  A Multiplatform Study of I/O Behavior on Petascale Supercomputers , 2015, HPDC.

[6]  Alok N. Choudhary,et al.  Improved parallel I/O via a two-phase run-time access strategy , 1993, CARN.

[7]  André Brinkmann,et al.  Improving Collective I/O Performance Using Non-volatile Memory Devices , 2016, 2016 IEEE International Conference on Cluster Computing (CLUSTER).

[8]  Satoshi Matsuoka,et al.  How file access patterns influence interference among cluster applications , 2014, 2014 IEEE International Conference on Cluster Computing (CLUSTER).

[9]  村井 均,et al.  NAS Parallel Benchmarks によるHPFの評価 , 2006 .

[10]  Rajeev Thakur,et al.  Pattern-Direct and Layout-Aware Replication Scheme for Parallel I/O Systems , 2013, 2013 IEEE 27th International Symposium on Parallel and Distributed Processing.

[11]  Brian Vinter,et al.  A Comparison of Three MPI Implementations , 2004 .

[12]  Robert B. Ross,et al.  On the role of burst buffers in leadership-class storage systems , 2012, 012 IEEE 28th Symposium on Mass Storage Systems and Technologies (MSST).

[13]  Hai Jin,et al.  Iteration Based Collective I/O Strategy for Parallel I/O Systems , 2014, 2014 14th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing.

[14]  Karsten Schwan,et al.  Six degrees of scientific data: reading patterns for extreme scale science IO , 2011, HPDC '11.

[15]  Nick Feltovich Nonparametric Tests of Differences in Medians: Comparison of the Wilcoxon–Mann–Whitney and Robust Rank-Order Tests , 2003 .

[16]  John Shalf,et al.  Using IOR to analyze the I/O Performance for HPC Platforms , 2007 .

[17]  Frank B. Schmuck,et al.  GPFS: A Shared-Disk File System for Large Computing Clusters , 2002, FAST.

[18]  Carlos Maltzahn,et al.  I/O acceleration with pattern detection , 2013, HPDC.

[19]  Robert Latham,et al.  Implementation and Evaluation of MPI Nonblocking Collective I/O , 2015, 2015 15th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing.

[20]  Dror G. Feitelson,et al.  Overview of the MPI-IO Parallel I/O Interface , 1996, Input/Output in Parallel and Distributed Computer Systems.

[21]  Ron Brightwell A Comparison of Three MPI Implementations for Red Storm , 2005, PVM/MPI.

[22]  O. J. Dunn Multiple Comparisons among Means , 1961 .