The Influence of Communication on the Performance of Co-allocation

In systems consistingof multiple clusters of processors interconnected by relatively slow connections such as our Distributed ASCI1 Supercomputer (DAS), jobs may request co-allocation, i.e., the simultaneous allocation of processors in different clusters. The performance of co-allocation may be severely impacted by the slowintercluster connections, and by the types of job requests. We distinguish different job request types ranging from ordered requests that specify the numbers of processors needed in each of the clusters, to flexible requests that only specify a total. We simulate multicluster systems with the FCFS policy-- and with two policies for placinga flexible request, one tries to balance cluster loads and one tries to fill clusters completely--to determine the response times under workloads consistingof a single or of different request types for different communication speeds across the intercluster connections. In addition to a synthetic workload, we also consider a workload derived from measurements of a real application on the DAS. We find that the communication speed difference has a severe impact on response times, that a relatively small amount of capacity is lost due to communication, and that for a mix of request types, the performance is determined not only by the separate behaviours of the different types of requests, but also by the way in which they interact.

[1]  Larry Rudolph,et al.  Towards Convergence in Job Schedulers for Parallel Supercomputers , 1996, JSSPP.

[2]  Tim Brecht,et al.  An Experimental Evaluation of Processor Pool-Based Scheduling for Shared-Memory NUMA Multiprocessors , 1997, JSSPP.

[3]  Mark J. Clement,et al.  The Performance Impact of Advance Reservation Meta-scheduling , 2000, JSSPP.

[4]  Vipin Kumar,et al.  Multi-capacity bin packing algorithms with applications to job scheduling under multiple constraints , 1999, Proceedings of the 1999 International Conference on Parallel Processing.

[5]  Dror G. Feitelson,et al.  Improved Utilization and Responsiveness with Gang Scheduling , 1997, JSSPP.

[6]  Henri E. Bal,et al.  MagPIe: MPI's collective communication operations for clustered wide area systems , 1999, PPoPP '99.

[7]  Uwe Schwiegelshohn,et al.  Theory and Practice in Parallel Job Scheduling , 1997, JSSPP.

[8]  Dror G. Feitelson,et al.  Packing Schemes for Gang Scheduling , 1996, JSSPP.

[9]  Henri E. Bal,et al.  Optimizing parallel applications for wide-area clusters , 1998, Proceedings of the First Merged International Parallel Processing Symposium and Symposium on Parallel and Distributed Processing.

[10]  Anca I. D. Bucur,et al.  The Influence of the Structure and Sizes of Jobs on the Performance of Co-allocation , 2000, JSSPP.

[11]  Kento Aida,et al.  Job Scheduling Scheme for Pure Space Sharing Among Rigid Jobs , 1998, JSSPP.

[12]  I. Deary,et al.  GLOBUS , 1989, The Lancet.

[13]  Viktor K. Prasanna,et al.  A framework for mapping with resource co-allocation in heterogeneous computing systems , 2000, Proceedings 9th Heterogeneous Computing Workshop (HCW 2000) (Cat. No.PR00556).

[14]  Ami Marowka,et al.  The GRID: Blueprint for a New Computing Infrastructure , 2000, Parallel Distributed Comput. Pract..

[15]  James Patton Jones,et al.  Scheduling for Parallel Supercomputing: A Historical Perspective of Achievable Utilization , 1999, JSSPP.

[16]  Peter M. A. Sloot,et al.  The distributed ASCI Supercomputer project , 2000, OPSR.