The Development of Parkbench and Performance Prediction

The authors present a brief overview of the development of benchmarks for parallel performance analysis and show how a set of widely accepted parallel benchmarks has emerged from the Parkbench initiative. At the lowest level, basic node parameters are captured by the LINPACK benchmark and the Genesis communications benchmarks, which give information on message-passing latency and bandwidth. At kernel and application levels, results are widely available for the NAS parallel benchmarks. A new release of the Genesis communications benchmarks is described, which addresses the important issue of the effect of memory hierarchy on message transfers. The rest of the paper is concerned with the way in which benchmark results can be used to predict performance of full applications. Three studies are considered: the PERFORM estimation tool and WHITEBOX benchmarking are research projects, indicating some possible directions for progress, while INTREPID is the basis of a commercial system for scheduling meta-applications based on performance models for the components.

[1]  Jack J. Dongarra,et al.  Performance of various computers using standard linear equations software in a FORTRAN environment , 1988, CARN.

[2]  Geoffrey C. Fox,et al.  The Perfect Club Benchmarks: Effective Performance Evaluation of Supercomputers , 1989, Int. J. High Perform. Comput. Appl..

[3]  Aad J. van der Steen,et al.  Benchmarking the Silicon Graphics Origin2000 System , 2000 .

[4]  F. H. Mcmahon,et al.  The Livermore Fortran Kernels: A Computer Test of the Numerical Performance Range , 1986 .

[5]  Report,et al.  Public International Benchmarks for Parallel Computers , 1993 .

[6]  Brian A. Wichmann,et al.  A Synthetic Benchmark , 1976, Comput. J..

[7]  Andrew Davison,et al.  Twelve Ways to Fool the Masses When Giving Performance Results on Parallel Computers , 1995 .

[8]  Roger W. Hockney The science of computer benchmarking , 1995, Software, environments, tools.

[9]  Mary K. Vernon,et al.  Poems: end-to-end performance design of large parallel adaptive computational systems , 1998, WOSP '98.

[10]  Rudolf Eigenmann,et al.  Benchmarking with real industrial applications: the SPEC High-Performance Group , 1996 .

[11]  Ami Marowka,et al.  The GRID: Blueprint for a New Computing Infrastructure , 2000, Parallel Distributed Comput. Pract..

[12]  Aad J. van der Steen The benchmark of the EuroBen group , 1991, Parallel Comput..

[13]  A J G Hey,et al.  PERFORM - A fast simulator for estimating program execution time , 1997 .

[14]  Anthony J. G. Hey,et al.  The Genesis distributed memory benchmarks , 1991, Parallel Comput..

[15]  David A. Padua,et al.  On the Automatic Parallelization of the Perfect Benchmarks , 1998, IEEE Trans. Parallel Distributed Syst..

[16]  Jack J. Dongarra,et al.  Computer benchmarks , 1993 .

[17]  David H. Bailey,et al.  Twelve ways to fool the masses when giving performance results on parallel computers , 1991 .

[18]  Anthony J. G. Hey,et al.  White-Box Benchmarking , 1998, Euro-Par.

[19]  David H. Bailey,et al.  The Nas Parallel Benchmarks , 1991, Int. J. High Perform. Comput. Appl..