Transitioning from file-based HPC workflows to streaming data pipelines with openPMD and ADIOS2

This paper aims to create a transition path from file-based IO to streaming-based workflows for scientific applications in an HPC environment. By using the openPMP-api, traditional workflows limited by filesystem bottlenecks can be overcome and flexibly extended for in situ analysis. The openPMD-api is a library for the description of scientific data according to the Open Standard for Particle-Mesh Data (openPMD). Its approach towards recent challenges posed by hardware heterogeneity lies in the decoupling of data description in domain sciences, such as plasma physics simulations, from concrete implementations in hardware and IO. The streaming backend is provided by the ADIOS2 framework, developed at Oak Ridge National Laboratory. This paper surveys two openPMD-based loosely-coupled setups to demonstrate flexible applicability and to evaluate performance. In loose coupling, as opposed to tight coupling, two (or more) applications are executed separately, e.g. in individual MPI contexts, yet cooperate by exchanging data. This way, a streaming-based workflow allows for standalone codes instead of tightlycoupled plugins, using a unified streaming-aware API and leveraging high-speed communication infrastructure available in modern compute clusters for massive data exchange. We determine new challenges in resource allocation and in the need of strategies for a flexible data distribution, demonstrating their influence on efficiency and scaling on the Summit compute system. The presented setups show the potential for a more flexible use of compute resources brought by streaming IO as well as the ability to increase throughput by avoiding filesystem bottlenecks.

[1]  Scott Klasky,et al.  On the Scalability of Data Reduction Techniques in Current and Upcoming HPC Systems from an Application Perspective , 2017, ISC Workshops.

[2]  V. Favre-Nicolin,et al.  Fast computation of scattering maps of nanostructures using graphical processing units , 2010, 1010.2641.

[3]  S. Luo,et al.  GAPD: a GPU-accelerated atom-based polychromatic diffraction simulation code. , 2018, Journal of synchrotron radiation.

[4]  E. Wes Bethel,et al.  Improving Performance of M-to-N Processing and Data Redistribution in In Transit Analysis and Visualization , 2020, EGPGV@Eurographics/EuroVis.

[5]  A. Mancuso,et al.  SIMEX: Simulation of Experiments at Advanced Light Sources , 2016, 1610.05980.

[6]  Arie Shoshani,et al.  Hello ADIOS: the challenges and lessons of developing leadership class I/O frameworks , 2014, Concurr. Comput. Pract. Exp..

[7]  James P. Ahrens,et al.  The ALPINE In Situ Infrastructure: Ascending from the Ashes of Strawman , 2017, ISAV@SC.

[8]  Martin Schulz,et al.  Flux: A Next-Generation Resource Management Framework for Large HPC Centers , 2014, 2014 43rd International Conference on Parallel Processing Workshops.

[9]  Erik Schultes,et al.  The FAIR Guiding Principles for scientific data management and stewardship , 2016, Scientific Data.

[10]  Kesheng Wu,et al.  Improving I/O Performance for Exascale Applications Through Online Data Layout Reorganization , 2021, IEEE Transactions on Parallel and Distributed Systems.

[11]  A. Buzmakov,et al.  Simulations of ultrafast x–ray laser experiments , 2017, Optics + Optoelectronics.

[12]  Axel Huebl,et al.  C++ & Python API for Scientific I/O with openPMD , 2019 .

[13]  Scott Klasky,et al.  Extending the Publish/Subscribe Abstraction for High-Performance I/O and Data Management at Extreme Scale , 2020, IEEE Data Eng. Bull..

[14]  Axel Hübl PIConGPU: Predictive Simulations of Laser-Particle Accelerators with Manycore Hardware , 2019 .

[15]  Jacek Krzywinski,et al.  Lightsource Unified Modeling Environment (LUME), a Start-to-End Simulation Ecosystem , 2021 .

[16]  Bronis R. de Supinski,et al.  The Design, Deployment, and Evaluation of the CORAL Pre-Exascale Systems , 2018, SC18: International Conference for High Performance Computing, Networking, Storage and Analysis.

[17]  Keichi Takahashi,et al.  ADIOS 2: The Adaptable Input Output System. A framework for high-performance data management , 2020, SoftwareX.

[18]  David S. Johnson,et al.  Near-optimal bin packing algorithms , 1973 .

[19]  SLADS: a parallel code for direct simulations of scattering of large anisotropic dense nanoparticle systems , 2017 .

[20]  Karsten Schwan,et al.  DataStager: scalable data staging services for petascale applications , 2009, HPDC '09.

[21]  Scott Klasky,et al.  Comparing Time-to-Solution for In Situ Visualization Paradigms at Scale , 2020, 2020 IEEE 10th Symposium on Large Data Analysis and Visualization (LDAV).