Infiniband scalability in Open MPI

Infiniband is becoming an important interconnect technology in high performance computing. Efforts in large scale Infiniband deployments are raising scalability questions in the HPC community. Open MPI, a new open source implementation of the MPI standard targeted for production computing, provides several mechanisms to enhance Infiniband scalability. Initial comparisons with MVAPICH, the most widely used Infiniband MPI implementation, show similar performance but with much better scalability characteristics. Specifically, small message latency is improved by up to 10% in medium/large jobs and memory usage per host is reduced by as much as 300%. In addition, Open MPI provides predictable latency that is close to optimal without sacrificing bandwidth performance

[1]  Ronald Minnich,et al.  A Network-Failure-Tolerant Message-Passing System for Terascale Clusters , 2002, ICS '02.

[2]  Anthony Skjellum,et al.  A High-Performance, Portable Implementation of the MPI Message Passing Interface Standard , 1996, Parallel Comput..

[3]  Fabrizio Petrini,et al.  QsNetII: An interconnect for supercomputing applica-tions , 2003 .

[4]  George Bosilca,et al.  Open MPI's TEG Point-to-Point Communications Methodology: Comparison to Existing Implementations , 2004, PVM/MPI.

[5]  S. Pakin,et al.  VMI 2 . 0 : A Dynamically Reconfigurable Messaging Layer for Availability , Usability , and Management , 2002 .

[6]  Ron Brightwell,et al.  The Portals 3.0 Message Passing Interface Revision 1.0 , 1999 .

[7]  Anthony Skjellum,et al.  Programming the InfiniBand Network Architecture for High Performance Message Passing Systems , 2003, ISCA PDCS.

[8]  Keith D. Underwood,et al.  A comparison of 4X InfiniBand and Quadrics Elan-4 technologies , 2004, 2004 IEEE International Conference on Cluster Computing (IEEE Cat. No.04EX935).

[9]  Dhabaleswar K. Panda,et al.  High Performance RDMA-Based MPI Implementation over InfiniBand , 2003, ICS '03.

[10]  Michael M. Resch,et al.  Towards Efficient Execution of MPI Applications on the Grid: Porting and Optimization Issues , 2003, Journal of Grid Computing.

[11]  Corporate The MPI Forum,et al.  MPI: a message passing interface , 1993, Supercomputing '93.

[12]  George Bosilca,et al.  Open MPI: Goals, Concept, and Design of a Next Generation MPI Implementation , 2004, PVM/MPI.

[13]  SkjellumAnthony,et al.  A high-performance, portable implementation of the MPI message passing interface standard , 1996 .

[14]  Ronald B. Brightwell,et al.  Scalability limitations of VIA-based technologies in supporting MPI , 2000 .

[15]  Jeffrey M. Squyres,et al.  The Component Architecture of Open MPI: Enabling Third-Party Collective Algorithms* , 2005 .

[16]  Andrew Lumsdaine,et al.  A Component Architecture for LAM/MPI , 2003, PVM/MPI.

[17]  Jack J. Dongarra,et al.  HARNESS and fault tolerant MPI , 2001, Parallel Comput..

[18]  Ron Brightwell A New MPI Implementation for Cray SHMEM , 2004, PVM/MPI.