NIOBE: An Intelligent I/O Bridging Engine for Complex and Distributed Workflows

In the age of data-driven computing, integrating High Performance Computing(HPC) and Big Data(BD) environments may be the key to increasing productivity and to driving scientific discovery forward. Scientific workflows consist of diverse applications (i.e., HPC simulations and BD analysis) each with distinct representations of data that introduce a semantic barrier between the two environments. To solve scientific problems at scale, accessing semantically different data from different storage resources is the biggest unsolved challenge. In this work, we aim to address a critical question: ”How can we exploit the existing resources and efficiently provide transparent access to data from/to both environments”. We propose iNtelligent I/O Bridging Engine(NIOBE), a new data integration framework that enables integrated data access for scientific workflows with asynchronous I/O and data aggregation. NIOBE performs the data integration using available I/O resources, in contrast to existing optimizations that ignore the I/O nodes present on the data path. In NIOBE, data access is optimized to consider both the ongoing production and the consumption of the data in the future. Experimental results show that with NIOBE, an integrated scientific workflow can be accelerated by up to 10x when compared to a no-integration baseline and by up to 133% compared to other state-of-the-art integration solutions.

[1]  Robert B. Ross,et al.  On the duality of data-intensive file system design: Reconciling HDFS and PVFS , 2011, 2011 International Conference for High Performance Computing, Networking, Storage and Analysis (SC).

[2]  Rajeev Thakur,et al.  Rethinking key–value store for parallel I/O optimization , 2014, 2014 International Workshop on Data Intensive Scalable Computing Systems.

[3]  Xian-He Sun,et al.  IRIS: I/O Redirection via Integrated Storage , 2018, ICS.

[4]  Frank B. Schmuck,et al.  GPFS: A Shared-Disk File System for Large Computing Clusters , 2002, FAST.

[5]  Scott Klasky,et al.  DataSpaces: an interaction and coordination framework for coupled simulation workflows , 2010, HPDC.

[6]  Kamil Iskra,et al.  ZOID: I/O-forwarding infrastructure for petascale architectures , 2008, PPoPP.

[7]  Jack J. Dongarra,et al.  Exascale computing and big data , 2015, Commun. ACM.

[8]  Russ Rew,et al.  NetCDF: an interface for scientific data access , 1990, IEEE Computer Graphics and Applications.

[9]  Karsten Schwan,et al.  DataStager: scalable data staging services for petascale applications , 2009, HPDC '09.

[10]  Kai Ren,et al.  IndexFS: Scaling File System Metadata Performance with Stateless Caching and Bulk Insertion , 2014, SC14: International Conference for High Performance Computing, Networking, Storage and Analysis.

[11]  Wei-keng Liao,et al.  Dynamically adapting file domain partitioning methods for collective I/O based on underlying parallel file system locking protocols , 2008, 2008 SC - International Conference for High Performance Computing, Networking, Storage and Analysis.

[12]  Robert Latham,et al.  Scalable I/O forwarding framework for high-performance computing systems , 2009, 2009 IEEE International Conference on Cluster Computing and Workshops.

[13]  Kai Ren,et al.  BatchFS: Scaling the File System Control Plane with Client-Funded Metadata Servers , 2014, 2014 9th Parallel Data Storage Workshop.

[14]  Scott Shenker,et al.  Spark: Cluster Computing with Working Sets , 2010, HotCloud.

[15]  Rajeev Thakur,et al.  Data sieving and collective I/O in ROMIO , 1998, Proceedings. Frontiers '99. Seventh Symposium on the Frontiers of Massively Parallel Computation.

[16]  Robert B. Ross,et al.  On the role of burst buffers in leadership-class storage systems , 2012, 012 IEEE 28th Symposium on Mass Storage Systems and Technologies (MSST).

[17]  Garrison Vaughan,et al.  Evolving Storage and Cyber Infrastructure at the NASA Center for Climate Simulation , 2018 .

[18]  Carlos Maltzahn,et al.  Ceph: a scalable, high-performance distributed file system , 2006, OSDI '06.

[19]  Rajeev Thakur,et al.  A Decoupled Execution Paradigm for Data-Intensive High-End Computing , 2012, 2012 IEEE International Conference on Cluster Computing.

[20]  Peter Desnoyers,et al.  Active flash: towards energy-efficient, in-situ data analytics on extreme-scale machines , 2013, FAST.

[21]  Allen D. Malony,et al.  Scaling Spark on HPC Systems , 2016, HPDC.

[22]  Yang Wang,et al.  Heterogeneity-Aware Collective I/O for Parallel I/O Systems with Hybrid HDD/SSD Servers , 2017, IEEE Transactions on Computers.

[23]  Gary Grider,et al.  MarFS, a Near-POSIX Interface to Cloud Objects , 2017, login Usenix Mag..

[24]  Emmanuel Jeannot,et al.  TAPIOCA: An I/O Library for Optimized Topology-Aware Data Aggregation on Large-Scale Supercomputers , 2017, 2017 IEEE International Conference on Cluster Computing (CLUSTER).

[25]  Surendra Byna,et al.  Boosting Application-Specific Parallel I/O Optimization Using IOSIG , 2012, 2012 12th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (ccgrid 2012).

[26]  Robert Latham,et al.  Parallel I/O in practice , 2006, SC.

[27]  Xian-He Sun,et al.  SciDP: Support HPC and Big Data Applications via Integrated Scientific Data Processing , 2018, 2018 IEEE International Conference on Cluster Computing (CLUSTER).

[28]  Scott Klasky,et al.  DataSpaces: an interaction and coordination framework for coupled simulation workflows , 2012, HPDC '10.

[29]  Zhao Zhang,et al.  FusionFS : a distributed file system for large scale data-intensive computing , 2013 .

[30]  Shujia Zhou,et al.  A lightweight scalable I/O utility for optimizing High-End Computing applications , 2008, 2008 IEEE International Symposium on Parallel and Distributed Processing.