In this paper, we describe Bridges, a new HPC resource that will integrate advanced memory technologies with a uniquely flexible, user-focused, data-centric environment to empower new research communities, bring desktop convenience to HPC, connect to campuses, and drive complex workflows. Bridges will differ from traditional HPC systems and support new communities through extensive interactivity, gateways (convenient web interfaces that hide complex functionality and ease access to HPC resources) and tools for gateway building, persistent databases and web servers, high-productivity programming languages, and virtualization. Bridges will feature three tiers of processing nodes having 128GB, 3TB, and 12TB of hardware-enabled coherent shared memory per node to support memory-intensive applications and ease of use, together with persistent database and web nodes and nodes for logins, data transfer, and system management. State-of-the-art Intel® Xeon® CPUs and NVIDIA Tesla GPUs will power Bridges' compute nodes. Multiple filesystems will provide optimal handling for different data needs: a high-performance, parallel, shared filesystem, node-local filesystems, and memory filesystems. Bridges' nodes and parallel filesystem will be interconnected by the Intel Omni-Path Fabric, configured in a topology developed by PSC to be optimal for the anticipated data-centric workload. Bridges will be a resource on XSEDE, the NSF Extreme Science and Engineering Discovery Environment, and will interoperate with other advanced cyberinfrastructure resources. Through a pilot project with Temple University, Bridges will develop infrastructure and processes for campus bridging, consisting of offloading jobs at periods of unusually high load to the other site and facilitating cross-site data management. Education, training, and outreach activities will raise awareness of Bridges and data-intensive science across K-12 and university communities, industry, and the general public.
[1]
Clark Glymour,et al.
Non-Gaussian methods and high-pass filters in the estimation of effective connections
,
2014,
NeuroImage.
[2]
Kenneth R. Koedinger,et al.
A Data Repository for the EDM Community: The PSLC DataShop
,
2010
.
[3]
SkjellumAnthony,et al.
A high-performance, portable implementation of the MPI message passing interface standard
,
1996
.
[4]
Mykola Pechenizkiy,et al.
Handbook of Educational Data Mining
,
2010
.
[5]
Nancy Wilkins-Diehr,et al.
Gateways to Discovery: Cyberinfrastructure for the Long Tail of Science
,
2014,
XSEDE '14.
[6]
Ian T. Foster,et al.
Jetstream: a self-provisioned, scalable science and engineering cloud environment
,
2015,
XSEDE.
[7]
P. Nowoczynski,et al.
Zest Checkpoint storage system for large supercomputers
,
2008,
2008 3rd Petascale Data Storage Workshop.
[8]
Todd R. Hunter,et al.
The Green Bank Telescope
,
2009,
Proceedings of the IEEE.
[9]
Matthew L. Jockers.
Macroanalysis: Digital Methods and Literary History
,
2013
.
[10]
Anthony Skjellum,et al.
Programming the InfiniBand Network Architecture for High Performance Message Passing Systems
,
2003,
ISCA PDCS.
[11]
Joel Welling,et al.
Blacklight: Coherent Shared Memory for Enabling Science
,
2017
.
[12]
Jared Yanovich,et al.
The data supercell
,
2012,
XSEDE '12.