Open MPI is a widely used open-source implementation of the MPI-2 standard that supports a variety of platforms and interconnects. Current versions of Open MPI, however, lack support for the Cray XE6 and XK6 architectures -- both of which use the Gemini System Interconnect. In this paper, we present extensions to natively support these architectures within Open MPI, describe and propose solutions for performance and scalability bottlenecks, and provide an extensive evaluation of our implementation, which is the first completely open-source MPI implementation for the Cray XE/XK system families used at 49,152 processes. Application and micro-benchmark results show that the performance and scaling characteristics of our implementation are similar to the vendor-supplied MPI's. Micro-benchmark results show short-data 1-byte and 1,024-byte message latencies of 1.20 μs and 4.13 μs, which are 10.00% and 39.71% better than the vendor-supplied MPI's, respectively. Our implementation achieves a bandwidth of 5.32 GB/s at 8 MB, which is similar to the vendor-supplied MPI's bandwidth at the same message size. Two Sequoia benchmark applications, LAMMPS and AMG2006, were also chosen to evaluate our implementation at scales up to 49,152 cores -- where we exhibited similar performance and scaling characteristics when compared to the vendor-supplied MPI implementation. LAMMPS achieved a parallel efficiency of 88.20% at 49,152 cores using Open MPI, which is on par with the vendor-supplied MPI's achieved parallel efficiency.
[1]
Darius Buntinas,et al.
A uGNI-Based MPICH2 Nemesis Network Module for the Cray XE
,
2011,
EuroMPI.
[2]
Nathan T. Hjelm,et al.
Open MPI for Cray XE/XK Systems
,
2012
.
[3]
Steve Plimpton,et al.
Fast parallel algorithms for short-range molecular dynamics
,
1993
.
[4]
Armin R. Mikler,et al.
NetPIPE: A Network Protocol Independent Performance Evaluator
,
1996
.
[5]
Brian W. Barrett,et al.
An Evaluation of Open MPI's Matching Transport Layer on the Cray XT
,
2007,
PVM/MPI.
[6]
Guillaume Mercier,et al.
Design and evaluation of Nemesis, a scalable, low-latency, message-passing communication subsystem
,
2006,
Sixth IEEE International Symposium on Cluster Computing and the Grid (CCGRID'06).
[7]
George Bosilca,et al.
Open MPI: Goals, Concept, and Design of a Next Generation MPI Implementation
,
2004,
PVM/MPI.
[8]
Larry Kaplan,et al.
The Gemini System Interconnect
,
2010,
2010 18th IEEE Symposium on High Performance Interconnects.