CARS: A contention-aware scheduler for efficient resource management of HPC storage systems

Abstract Many scientific applications are becoming more and more data intensive. As the data volume continues to grow, the data movement between storage and compute nodes has turned into a crucial performance bottleneck for many data-intensive applications. Burst buffer provides a promising solution for these applications by absorbing bursty I/O traffic. However, the resource allocation and management strategies for burst buffer are not well studied. The existing bandwidth based strategies may cause severe I/O contention when a large number of I/O-intensive jobs access the burst buffer system concurrently. In this study, we present a contention-aware resource scheduling (CARS) strategy to manage the burst buffer resources and coordinate concurrent data-intensive jobs. The experimental results show that the proposed CARS framework outperforms the existing allocation strategies and improves both the job performance and the system utilization.

[1]  Teng Wang,et al.  An Ephemeral Burst-Buffer File System for Scientific Applications , 2016, SC16: International Conference for High Performance Computing, Networking, Storage and Analysis.

[2]  Marianne Winslett,et al.  A Multiplatform Study of I/O Behavior on Petascale Supercomputers , 2015, HPDC.

[3]  Karsten Schwan,et al.  Managing Variability in the IO Performance of Petascale Storage Systems , 2010, 2010 ACM/IEEE International Conference for High Performance Computing, Networking, Storage and Analysis.

[4]  Robert Latham,et al.  Scalable I/O and analytics , 2009 .

[5]  Yong Chen,et al.  Contention-Aware Resource Scheduling for Burst Buffer Systems , 2018, ICPP Workshops.

[6]  Satoshi Matsuoka,et al.  A User-Level InfiniBand-Based File System and Checkpoint Strategy for Burst Buffers , 2014, 2014 14th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing.

[7]  Teng Wang,et al.  TRIO: Burst Buffer Based I/O Orchestration , 2015, 2015 IEEE International Conference on Cluster Computing.

[8]  Bo Peng,et al.  Five-Hundred-Meter Aperture Spherical Telescope Project , 2001 .

[9]  Franck Cappello,et al.  Scheduling the I/O of HPC Applications Under Congestion , 2015, 2015 IEEE International Parallel and Distributed Processing Symposium.

[10]  Carlos Maltzahn,et al.  DAOS and Friends: A Proposal for an Exascale Storage System , 2016, SC16: International Conference for High Performance Computing, Networking, Storage and Analysis.

[11]  Purushotham Bangalore,et al.  Managing I/O Interference in a Shared Burst Buffer System , 2016, 2016 45th International Conference on Parallel Processing (ICPP).

[12]  Xian-He Sun,et al.  Harmonia: An Interference-Aware Dynamic I/O Scheduler for Shared Non-volatile Burst Buffers , 2018, 2018 IEEE International Conference on Cluster Computing (CLUSTER).

[13]  Bronis R. de Supinski,et al.  The Design, Deployment, and Evaluation of the CORAL Pre-Exascale Systems , 2018, SC18: International Conference for High Performance Computing, Networking, Storage and Analysis.

[14]  John Shalf,et al.  The International Exascale Software Project roadmap , 2011, Int. J. High Perform. Comput. Appl..

[15]  Jia Wang,et al.  I/O-Aware Batch Scheduling for Petascale Computing Systems , 2015, 2015 IEEE International Conference on Cluster Computing.

[16]  Soonwook Hwang,et al.  Accelerating a Burst Buffer Via User-Level I/O Isolation , 2017, 2017 IEEE International Conference on Cluster Computing (CLUSTER).

[17]  Robert B. Ross,et al.  On the role of burst buffers in leadership-class storage systems , 2012, 012 IEEE 28th Symposium on Mass Storage Systems and Technologies (MSST).

[18]  Dong H. Ahn,et al.  Scalable I/O-Aware Job Scheduling for Burst Buffer Enabled HPC Clusters , 2016, HPDC.

[19]  Sorin Faibish,et al.  Jitter-free co-processing on a prototype exascale storage stack , 2012, 012 IEEE 28th Symposium on Mass Storage Systems and Technologies (MSST).

[20]  Michael Lang,et al.  Active Burst-Buffer: In-Transit Processing Integrated into Hierarchical Storage , 2016, 2016 IEEE International Conference on Networking, Architecture and Storage (NAS).

[21]  Adrien Lèbre,et al.  I/O Scheduling Service for Multi-Application Clusters , 2006, 2006 IEEE International Conference on Cluster Computing.

[22]  John Shalf,et al.  Characterizing and predicting the I/O performance of HPC applications using a parameterized synthetic benchmark , 2008, 2008 SC - International Conference for High Performance Computing, Networking, Storage and Analysis.

[23]  Song Jiang,et al.  IOrchestrator: Improving the Performance of Multi-node I/O Systems via Inter-Server Coordination , 2010, 2010 ACM/IEEE International Conference for High Performance Computing, Networking, Storage and Analysis.

[24]  Karsten Schwan,et al.  Six degrees of scientific data: reading patterns for extreme scale science IO , 2011, HPDC '11.

[25]  Surendra Byna,et al.  Accelerating Science with the NERSC Burst Buffer Early User Program , 2016 .