0% found this document useful (0 votes)
197 views

Project Stargate

Project StarGate aimed to connect supercomputing resources across several national laboratories to enable remote visualization and analysis of extremely large cosmological simulation data. It used high-speed networks and OptIPortals to stream petabyte-scale simulation output from Oak Ridge to graphics clusters at Argonne and San Diego for interactive rendering and display. The project demonstrated that computational resources do not need to be isolated, allowing scientists to analyze results from where best suited without data being trapped locally.

Uploaded by

geekboy
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
197 views

Project Stargate

Project StarGate aimed to connect supercomputing resources across several national laboratories to enable remote visualization and analysis of extremely large cosmological simulation data. It used high-speed networks and OptIPortals to stream petabyte-scale simulation output from Oak Ridge to graphics clusters at Argonne and San Diego for interactive rendering and display. The project demonstrated that computational resources do not need to be isolated, allowing scientists to analyze results from where best suited without data being trapped locally.

Uploaded by

geekboy
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 14

Project StarGate

ANL * Calit2 * LBNL * NICS * ORNL * SDSC


Credits
Argonne National Laboratory Lawrence Berkeley National
Network/Systems Laboratory (ESnet)
 Linda Winkler  Eli Dart
 Loren Jan Wilson
National Institute for
Visualization Computational Sciences
 Joseph Insley  Nathaniel Mendoza
 Eric Olsen
 Mark Hereld Oak Ridge National Laboratory
 Michael Papka  Susan Hicks

Calit2@UCSD San Diego Supercomputer Center


 Larry Smarr (Overall Concept) Science application
 Brian Dunne (Networking)  Michael Norman
 Joe Keefe (OptIPortal)  Rick Wagner (coordinator)
 Kai Doerr, Falko Kuester Network
(CGLX)  Tom Hutton

• ANL * Calit2 * LBNL * NICS * ORNL * SDSC


Exploring Cosmology With Supercomputers,
Supernetworks, and Supervisualization
Intergalactic medium on 2 Glyr scale
• 40963 particle/cell
hydrodynamic
cosmology simulation
• NICS Kraken (XT5)
– 16,384 cores
• Output
– 148 TB movie output
(0.25 TB/file)
– 80 TB diagnostic
dumps (8 TB/file)
Science: Norman, Harkness,Paschos SDSC
Visualization: Insley, ANL; Wagner SDSC
• ANL * Calit2 * LBNL * NICS * ORNL * SDSC
Project StarGate Goals
• Explore Use of OptIPortals as
Petascale Supercomputer
OptIPortal@SDSC
“Scalable Workstations”
• Exploit Dynamic 10 Gbs
Circuits on ESnet
• Connect Hardware Resources
at ORNL, ANL, SDSC
• Show that Data Need Not be
Trapped by the Network Rick Wagner Mike Norman
“Event Horizon”

• ANL * Calit2 * LBNL * NICS * ORNL * SDSC


Why Supercomputer Centers Shouldn’t Be
Data Black Holes or Island Universes

• Results are the


Intellectual Property of the
Investigator, Not the Center
Where it was Computed
• Petascale HPC Machines
Not Ideal for Analysis/Viz
• Doesn’t Take Advantage of
Local CI Resources on
Campuses (e.g., Triton) or at
other National Facilities
(e.g., ANL Eureka)

• ANL * Calit2 * LBNL * NICS * ORNL * SDSC


Opening Up 10Gbps Data Path
ORNL/NICS to ANL to SDSC

End-to-End Coupling of User with DOE/NSF HPC Facilities


StarGate Network & Hardware
ALCF
ESnet
DOE Eureka
Science Data Network (SDN) 100 Dual Quad Core Xeon Servers
> 10 Gb/s fiber optic network 200 NVIDIA Quadro FX GPUs in 50
Dynamic VLANs configured Quadro Plex S4 1U enclosures
using OSCARS 3.2 TB RAM rendering

SDSC

Challenge: Kraken
is not on ESnet NICS

visualization NSF TeraGrid Kraken simulation


Cray XT5
8,256 Compute Nodes
Calit2/SDSC OptIPortal1
99,072 Compute Cores
20 30” (2560 x 1600 pixel) LCD panels
129 TB RAM
10 NVIDIA Quadro FX 4600 graphics
cards > 80 gigapixels
10 Gb/s network throughout
• ANL * Calit2 * LBNL * NICS * ORNL * SDSC
ESnet StarGate Streaming Rendering
3 A media bridge at the border provides secure
ALCF Internal access to the parallel rendering streams.

gs1.intrepid.alcf.anl.gov

SDSC ALCF

Updated instructions are


5 sent back to the renderer to
change views, or load a
different dataset.

The full image is broken into


subsets (tiles). The tiles are
flPy, a parallel (MPI) tiled image/movie continuously encoded as a
4 viewer composites the individual movies, separate movies.
and synchronizes the movie playback across
the OptIPortal rendering nodes.
Simulation volume is rendered using vl3 , a
1 parallel (MPI) volume renderer utilizing
Eureka’s GPUs. The rendering changes
views steadily to highlight 3D structure.

• ANL * Calit2 * LBNL * NICS * ORNL * SDSC


Test animation of 1/64 of the data volume
(10243 region)

www.mcs.anl.gov/~insley/ENZO/BAO/B4096/enzo-b4096-1024subregion-test.mov
• ANL * Calit2 * LBNL * NICS * ORNL * SDSC
Data Moved
• ORNL to ANL data transfer nodes
• 577 time steps ~148TB
• Peak bandwidth ~2.4Gb/s
• Disk to disk
• GridFTP, Multiple Simultaneous Transfers, Each with Multiple TCP Connects
• Average Aggregate Bandwidth <800mb/s, Using Multiple Transfers

• Additionally
• Pre-Transfer: Data was Stored in ORNL HPSS, Had to be
Staged to Disk on Data Transfer Nodes
• One Moved to HPSS Partition, Cant Move Data Back
• Post-Transfer: Each Time Step was a Tar File, Had to Untar
• Moving Forward, will Need Direct High-Bandwidth Path
from Kraken (NICS) to Eureka (ALCF)

• ANL * Calit2 * LBNL * NICS * ORNL * SDSC


ANL Eureka Graphics Cluster
• Data Analytics and Visualization Cluster at ALCF
• (2) Head Nodes, (100) Compute Nodes
• (2) Nvidia Quadro FX5600 Graphics Cards
• (2) XEON E5405 2.00 GHz Quad Core Processors
• 32 GB RAM: (8) 4 Rank, 4GB DIMMS
• (1) Myricom 10G CX4 NIC
• (2) 250GB Local Disks; (1) System, (1) Minimal Scratch
• 32 GFlops per Server

• ANL * Calit2 * LBNL * NICS * ORNL * SDSC


Visualization Pipeline

• vl3 – Hardware Accelerated Volume Rendering Library


• 40963 Volume on 65 Nodes of Eureka
• Enzo Reader can Load from Native HDF5 Format
• Uniform Grid and AMR, Resampled to Uniform grid
• Locally Run Interactively on Subset of Data
• On a Local Workstation, 5123 Subvolume
• Batch for Generating Animations on Eureka
• Working Toward Remote Display and Control

• ANL * Calit2 * LBNL * NICS * ORNL * SDSC


vl3 Rendering
Performance on Eureka

• Image Size: 4096x4096


• Number of Samples: 4096

Data Size Number of Load Time Render/Composite


Processors/ Time
Graphics Cards
20483 17 2min 27sec 9.22 sec
40963 129 5min 10sec 4.51 sec
64003 (AMR) 129 4min 17sec 13.42sec

Note Data I/O Bottleneck

• ANL * Calit2 * LBNL * NICS * ORNL * SDSC


Next Experiments

• SC09 - Stream a 4Kx2K Movie From ANL Storage Device


to OptIPortable on Show Floor

• Mike Norman is a 2009 INCITE investigator


– 6 M SU on Jaguar
– Supersonic MHD Turbulence Simulations for Star Formation
– Use Similar Data Path for This to Show Replicability

• Can DOE Make This New Mode Available to Other Users?

• ANL * Calit2 * LBNL * NICS * ORNL * SDSC

You might also like