Plasma fusion code coupling using scalable I/O services and scientific workflows

In order to understand the complex physics of mother nature, physicist often use many approximations to understand one area of physics and then write a simulation to reduce these equations to ones that can be solved on a computer. Different approximations lead to different equations that model different physics, which can often lead to a completely different simulation code. As computers become more powerful, scientists can either write one simulation that models all of the physics or they produce several codes each for different portions of the physics and then 'couple' these codes together. In this paper, we concentrate on the latter, where we look at our code coupling approach for modeling a full device fusion reactor. There are many approaches to code coupling. Our first approach was using Kepler workflows to loosely couple three codes via files (memory-to-disk-to-memory coupling). This paper describes our new approach moving towards using memory-to-memory data exchange to allow for a tighter coupling. Our approach focuses on a method which brings together scientific workflows along with staging I/O methods for code coupling. Staging methods use additional compute nodes to perform additional tasks such as data analysis, visualization, and NxM transfers for code coupling. In order to transparently allow application scientist to switch from memory to memory coupling to memory to disk to memory coupling, we have been developing a framework that can switch between these two I/O methods and then automate other workflow tasks. Our hybrid approach allows application scientist to easily switch between in-memory coupling and file-based coupling on-the-fly, which aids debugging these complex configurations.

[1]  Scott Klasky,et al.  Tracking Files in the Kepler Provenance Framework , 2009, SSDBM.

[2]  Scott Klasky,et al.  Plasma Edge Kinetic-MHD Modeling in Tokamaks Using Kepler Workflow for Code Coupling, Data Management and Visualization , 2008 .

[3]  Karsten Schwan,et al.  Extending I/O through high performance data services , 2009, 2009 IEEE International Conference on Cluster Computing and Workshops.

[4]  Scott Klasky,et al.  Workflow automation for processing plasma fusion simulation data , 2007, WORKS '07.

[5]  W. Park,et al.  Plasma simulation studies using multilevel physics models , 1999 .

[6]  Scott Klasky,et al.  DART: a substrate for high speed asynchronous data IO , 2008, HPDC '08.

[7]  Edward A. Lee,et al.  Scientific workflow management and the Kepler system , 2006, Concurr. Comput. Pract. Exp..

[8]  Arie Shoshani,et al.  EFFIS: An End-to-end Framework for Fusion Integrated Simulation , 2010, 2010 18th Euromicro Conference on Parallel, Distributed and Network-based Processing.

[9]  T. Hahm,et al.  Size Scaling of Turbulent Transport in Magnetically Confined Plasmas , 2002 .

[10]  Arie Shoshani,et al.  Toward a first-principles integrated simulation of tokamak edge plasmas , 2008 .

[11]  Karsten Schwan,et al.  Flexible IO and integration for scientific codes through the adaptable IO system (ADIOS) , 2008, CLADE '08.

[12]  Scott Klasky,et al.  Scientific Process Automation and Workflow Management , 2009, Scientific Data Management.

[13]  Scott Klasky,et al.  Collaboration portal for petascale simulations , 2009, 2009 International Symposium on Collaborative Technologies and Systems.

[14]  H. Wilson,et al.  Numerical studies of edge localized instabilities in tokamaks , 2002 .

[15]  Choong-Seock Chang,et al.  Numerical study of neoclassical plasma pedestal in a tokamak geometry , 2004 .

[16]  James Demmel,et al.  A Data Broker for Distributed Computing Environments , 2001, International Conference on Computational Science.

[17]  Li Zhang,et al.  A Dynamic Geometry-Based Shared Space Interaction Framework for Parallel Scientific Applications , 2004, HiPC.

[18]  Karsten Schwan,et al.  Adaptable, metadata rich IO methods for portable high performance IO , 2009, 2009 IEEE International Symposium on Parallel & Distributed Processing.

[19]  Scott Klasky,et al.  Petascale IO Using The Adaptable IO System , 2009 .