A flexible I/O arbitration framework for netCDF‐based big data processing workflows on high‐end supercomputers

On the verge of the convergence between high‐performance computing and Big Data processing, it has become increasingly prevalent to deploy large‐scale data analytics workloads on high‐end supercomputers. Such applications often come in the form of complex workflows with various different components, assimilating data from scientific simulations as well as from measurements streamed from sensor networks, such as radars and satellites. For example, as part of the Flagship 2020 (post‐K) supercomputer project of Japan, RIKEN is investigating the feasibility of a highly accurate weather forecasting system that would provide a real‐time outlook for severe guerrilla rainstorms. One of the main performance bottlenecks of this application is the lack of efficient communication among workflow components, which currently takes place over the parallel file system.In this paper, we present an initial study of a direct communication framework designed for complex workflows that eliminates unnecessary file I/O among components. Specifically, we propose an I/O arbitration layer that provides direct parallel data transfer (both synchronous and asynchronous) among job components that rely on the netCDF interface for performing I/O operations. Our solution requires only minimal modifications to application code. Moreover, we propose a configuration file–based approach that allows users to specify the desired data transfer pattern among workflow components, offering a general solution for different application contexts. We present a preliminary evaluation of the proposed framework on the K Computer (running on up to 4800 compute nodes) using RIKEN's experimental weather forecasting workflow as a case study.

[1]  Kevin W. Manning,et al.  The Integrated WRF/Urban Modeling System: Development, Evaluation, and Applications to Urban Environmental Problems , 2010 .

[2]  Yutaka Ishikawa,et al.  Toward a General I/O Arbitration Framework for netCDF Based Big Data Processing , 2016, Euro-Par.

[3]  Yutaka Ishikawa,et al.  Adaptive transport service selection for MPI with InfiniBand network , 2015, ExaMPI '15.

[4]  Jack J. Dongarra,et al.  Exascale computing and big data , 2015, Commun. ACM.

[5]  Mariana Vertenstein,et al.  A new flexible coupler for earth system modeling developed for CCSM4 and CESM1 , 2012, Int. J. High Perform. Comput. Appl..

[6]  Jay Walter Larson,et al.  The Model Coupling Toolkit: A New Fortran90 Toolkit for Building Multiphysics Parallel Coupled Models , 2005, Int. J. High Perform. Comput. Appl..

[7]  S. Wilson,et al.  A simple method for integrating a complex model into an ensemble data assimilation system using MPI , 2015, Environ. Model. Softw..

[8]  Yousuke Sato,et al.  Influence of grid aspect ratio on planetary boundary layer turbulence in large-eddy simulations , 2015 .

[9]  J. Dudhia,et al.  Coupling an Advanced Land Surface–Hydrology Model with the Penn State–NCAR MM5 Modeling System. Part I: Model Implementation and Sensitivity , 2001 .

[10]  Koji Terasaki,et al.  Big Ensemble Data Assimilation in Numerical Weather Prediction , 2015, Computer.

[11]  Yousuke Sato,et al.  Impacts of cloud microphysics on trade wind cumulus: which cloud microphysics processes contribute to the diversity in a large eddy simulation? , 2015, Progress in Earth and Planetary Science.

[12]  S. Valcke,et al.  The OASIS3 coupler: a European climate modelling community software , 2012 .

[13]  Justin M. Wozniak,et al.  Lessons Learned from Building In Situ Coupling Frameworks , 2015, ISAV@SC.

[14]  Robert Latham,et al.  Toward a General I/O Layer for Parallel-Visualization Applications , 2011, IEEE Computer Graphics and Applications.

[15]  L. Liu,et al.  C-Coupler1: a Chinese community coupler for Earth System Modelling , 2014 .

[16]  Litian Liu,et al.  Improving data transfer for model coupling , 2015 .

[17]  Cecelia DeLuca,et al.  Coupling technologies for Earth System Modelling , 2012 .

[18]  David Pozo-Vázquez,et al.  One-Day-Ahead Streamflow Forecasting Using Artificial Neural Networks and a Meteorological Mesoscale Model , 2015 .

[19]  Jeffrey L. Anderson,et al.  The Exchange Grid: A mechanism for data exchange between Earth System components on independent grids , 2006 .

[20]  Wolfgang Frings,et al.  Scalable Control and Monitoring of Supercomputer Applications Using an Integrated Tool Framework , 2011, 2011 40th International Conference on Parallel Processing Workshops.

[21]  Chris H. Q. Ding,et al.  CPL6: The New Extensible, High Performance Parallel Coupler for the Community Climate System Model , 2005, Int. J. High Perform. Comput. Appl..

[22]  Fumiyoshi Shoji,et al.  The K Computer , 2017 .

[23]  Karsten Schwan,et al.  Plasma fusion code coupling using scalable I/O services and scientific workflows , 2009, WORKS '09.

[24]  Fan Zhang,et al.  Enabling In-situ Execution of Coupled Scientific Workflow on Multi-core Platform , 2012, 2012 IEEE 26th International Parallel and Distributed Processing Symposium.

[25]  Graham D. Riley,et al.  Coupling integrated Earth System Model components with BFG2 , 2009, Concurr. Comput. Pract. Exp..

[26]  David C. Goodrich,et al.  An integrated modelling framework of catchment‐scale ecohydrological processes: 1. Model description and tests over an energy‐limited watershed , 2014 .

[27]  Istvan Szunyogh,et al.  Efficient data assimilation for spatiotemporal chaos: A local ensemble transform Kalman filter , 2005, physics/0511236.

[28]  Guangwen Yang,et al.  Interpolation oriented parallel communication to optimize coupling in earth system modeling , 2014, Frontiers of Computer Science.

[29]  Graham D. Riley,et al.  Sharing experiences and outlook on Coupling Technologies for Earth System Models , 2016 .