Investigation of leading HPC I/O performance using a scientific-application derived benchmark

With the exponential growth of high-fidelity sensor and simulated data, the scientific community is increasingly reliant on ultrascale HPC resources to handle their data analysis requirements. However, to utilize such extreme computing power effectively, the I/O components must be designed in a balanced fashion, as any architectural bottleneck will quickly render the platform intolerably inefficient. To understand I/O performance of data-intensive applications in realistic computational settings, we develop a lightweight, portable benchmark called MADbench2, which is derived directly from a large-scale Cosmic Microwave Background (CMB) data analysis package. Our study represents one of the most comprehensive I/O analyses of modern parallel filesystems, examining a broad range of system architectures and configurations, including Lustre on the Cray XT3 and Intel Itanium2 cluster; GPFS on IBM Power5 and AMD Opteron platforms; two BlueGene/L installations utilizing GPFS and PVFS2 filesystems; and CXFS on the SGI Altix3700. We present extensive synchronous I/O performance data comparing a number of key parameters including concurrency, POSIX- versus MPI-IO, and unique- versus shared-file accesses, using both the default environment as well as highly-tuned I/O parameters. Finally, we explore the potential of asynchronous I/O and quantify the volume of computation required to hide a given volume of I/O. Overall our study quantifies the vast differences in performance and functionality of parallel filesystems across state-of-the-art platforms, while providing system designers and computational scientists a lightweight tool for conducting further analyses.

[1]  Leonid Oliker,et al.  Integrated performance monitoring of a cosmology application on leading HEC platforms , 2005, 2005 International Conference on Parallel Processing (ICPP'05).

[2]  Leonid Oliker,et al.  Performance Characteristics of a Cosmology Package on Leading HPC Architectures , 2004, HiPC.

[3]  Daniel C. Duffy,et al.  Beyond the storage area network: data intensive computing in a distributed environment , 2005, 22nd IEEE / 13th NASA Goddard Conference on Mass Storage Systems and Technologies (MSST'05).

[4]  Frank B. Schmuck,et al.  GPFS: A Shared-Disk File System for Large Computing Clusters , 2002, FAST.

[5]  Subhash Saini,et al.  Parallel I/O Performance Characterization of Columbia and NEC SX-8 Superclusters , 2007, 2007 IEEE International Parallel and Distributed Processing Symposium.

[6]  Robert B. Ross,et al.  Efficient structured data access in parallel file systems , 2003, 2003 Proceedings IEEE International Conference on Cluster Computing.

[7]  Mahadev Satyanarayanan,et al.  Scale and performance in a distributed file system , 1987, SOSP '87.

[8]  Rolf Rabenseifner,et al.  Effective File-I/O Bandwidth Benchmark , 2000, Euro-Par.

[9]  John Shalf,et al.  Using IOR to analyze the I/O Performance for HPC Platforms , 2007 .

[10]  Robert B. Ross,et al.  Using MPI-2: Advanced Features of the Message Passing Interface , 2003, CLUSTER.

[11]  Julian Borrill MADCAP - The Microwave Anisotropy Dataset Computational Analysis Package , 1999 .

[12]  Brad Gallagher,et al.  Scientific Applications on the Massively Parallel BG/L Machine , 2006, PDPTA.

[13]  P. Braam Systems for Clusters from a Protocol Perspective , 1999 .

[14]  Robert Latham,et al.  A next-generation parallel file system for Linux cluster. , 2004 .