The goal of this research was to investigate the potential for employing dynamic, decentralized software architectures to achieve reliability in future high-performance computing platforms. These architectures, inspired by peer-to-peer networks such as botnets that already scale to millions of unreliable nodes, hold promise for enabling scientific applications to run usefully on next-generation exascale platforms ({approx} 10{sup 18} operations per second). Traditional parallel programming techniques suffer rapid deterioration of performance scaling with growing platform size, as the work of coping with increasingly frequent failures dominates over useful computation. Our studies suggest that new architectures, in which failures are treated as ubiquitous and their effects are considered as simply another controllable source of error in a scientific computation, can remove such obstacles to exascale computing for certain applications. We have developed a simulation framework, as well as a preliminary implementation in a large-scale emulation environment, for exploration of these 'fault-oblivious computing' approaches. High-performance computing (HPC) faces a fundamental problem of increasing total component failure rates due to increasing system sizes, which threaten to degrade system reliability to an unusable level by the time the exascale range is reached ({approx} 10{sup 18} operations per second, requiring of order millions of processors). As computermore » scientists seek a way to scale system software for next-generation exascale machines, it is worth considering peer-to-peer (P2P) architectures that are already capable of supporting 10{sup 6}-10{sup 7} unreliable nodes. Exascale platforms will require a different way of looking at systems and software because the machine will likely not be available in its entirety for a meaningful execution time. Realistic estimates of failure rates range from a few times per day to more than once per hour for these platforms. P2P architectures give us a starting point for crafting applications and system software for exascale. In the context of the Internet, P2P applications (e.g., file sharing, botnets) have already solved this problem for 10{sup 6}-10{sup 7} nodes. Usually based on a fractal distributed hash table structure, these systems have proven robust in practice to constant and unpredictable outages, failures, and even subversion. For example, a recent estimate of botnet turnover (i.e., the number of machines leaving and joining) is about 11% per week. Nonetheless, P2P networks remain effective despite these failures: The Conficker botnet has grown to {approx} 5 x 10{sup 6} peers. Unlike today's system software and applications, those for next-generation exascale machines cannot assume a static structure and, to be scalable over millions of nodes, must be decentralized. P2P architectures achieve both, and provide a promising model for 'fault-oblivious computing'. This project aimed to study the dynamics of P2P networks in the context of a design for exascale systems and applications. Having no single point of failure, the most successful P2P architectures are adaptive and self-organizing. While there has been some previous work applying P2P to message passing, little attention has been previously paid to the tightly coupled exascale domain. Typically, the per-node footprint of P2P systems is small, making them ideal for HPC use. The implementation on each peer node cooperates en masse to 'heal' disruptions rather than relying on a controlling 'master' node. Understanding this cooperative behavior from a complex systems viewpoint is essential to predicting useful environments for the inextricably unreliable exascale platforms of the future. We sought to obtain theoretical insight into the stability and large-scale behavior of candidate architectures, and to work toward leveraging Sandia's Emulytics platform to test promising candidates in a realistic (ultimately {ge} 10{sup 7} nodes) setting. Our primary example applications are drawn from linear algebra: a Jacobi relaxation solver for the heat equation, and the closely related technique of value iteration in optimization. We aimed to apply P2P concepts in designing implementations capable of surviving an unreliable machine of 10{sup 6} nodes.« less
[1]
Tang,et al.
Self-Organized Criticality: An Explanation of 1/f Noise
,
2011
.
[2]
Stuart A. Kauffman,et al.
The origins of order
,
1993
.
[3]
Ronald G. Minnich,et al.
Ten Million and One Penguins or Lessons Learned from booting millions of virtual machines on HPC systems.
,
2010
.
[4]
David S. Johnson,et al.
Computers and Intractability: A Guide to the Theory of NP-Completeness
,
1978
.
[5]
Lyn C. Thomas,et al.
Serial and parallel value iteration algorithms for discounted Markov decision processes
,
1993
.
[6]
Andrew G. Barto,et al.
Reinforcement learning
,
1998
.
[7]
Ümit V. Çatalyürek,et al.
Hypergraph-based Dynamic Load Balancing for Adaptive Scientific Computations
,
2007,
2007 IEEE International Parallel and Distributed Processing Symposium.
[8]
Andrew G. Barto,et al.
Convergence of Indirect Adaptive Asynchronous Value Iteration Algorithms
,
1993,
NIPS.
[9]
Ronald G. Minnich,et al.
Approaches for Scalable Modeling and Emulation of Cyber Systems: LDRD Final Report
,
2009
.
[10]
Stéphane Genaud,et al.
P2P-MPI: A Peer-to-Peer Framework for Robust Execution of Message Passing Parallel Programs on Grids
,
2007,
Journal of Grid Computing.
[11]
Richard S. Sutton,et al.
Dimensions of Reinforcement Learning
,
1998
.
[12]
Stephen Wolfram,et al.
A New Kind of Science
,
2003,
Artificial Life.
[13]
Jackson Mayo,et al.
Methodologies for advance warning of compute cluster problems via statistical analysis: a case study
,
2009,
Resilience '09.
[14]
John Daly.
A Model for Predicting the Optimum Checkpoint Interval for Restart Dumps
,
2003,
International Conference on Computational Science.