<<

Optimizing Network Usage on and Sierra

Bronis R. de Supinski Chief Technology Officer Livermore Compung June 23, 2016

LLNL-PRES-695482 This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under contract DE-AC52-07NA27344. Lawrence Livermore National Security, LLC My focus is NNSA ASC ATS plaorms at LLNL

Sequoia (LLNL)

ATS 1 – Trinity (LANL/SNL)

ATS 2 – Sierra (LLNL)

ATS 3 – Crossroads (LANL/SNL) Advanced Technology Technology Advanced (ATS) Systems System Delivery ATS 4 – (LLNL)

ATS 5 – (LANL/SNL)

Tri-lab Linux Capacity Cluster II (TLCC II)

Procure& CTS 1 Deploy Use CTS 2

Commodity Technology (CTS) Systems Retire

‘13 ‘14 ‘15 ‘16 ‘17 ‘18 ‘19 ‘20 ‘21 ‘22 ‘23 Fiscal Year Sequoia and Sierra are the current and next-generaon Advanced Technology Systems at LLNL

2 LLNL-PRES-695482 Sequoia provides previously unprecedented levels of capability and concurrency § Sequoia stascs — 20 petaFLOP/s peak — 17 petaFLOP/s LINPACK — Memory 1.5 PB, 4 PB/s bandwidth — 98,304 nodes — 1,572,864 cores — 3 PB/s link bandwidth — 60 TB/s bi-secon bandwidth — 0.5−1.0 TB/s Lustre bandwidth — 50 PB disk

§ 9.6MW power, 4,000 2

§ Third generaon IBM BlueGene

3 LLNL-PRES-695482 The BG/Q compute chip integrates processors, memory and networking logic into one chip

§ 16 user + 1 OS + 1 redundant cores — 4-way mul-threaded, 1.6 GHz 64-bit — 16kB/16kB L1 I/D caches — Quad FPUs (4-wide DP SIMD) — Peak: 204.8 GFLOPS @ 55 W

§ Shared 32 MB eDRAM L2 cache — Mulversioned cache

§ Dual memory controller — 16 GB DDR3 memory (1.33 Gb/s) — 2 * 16 byte-wide interface (+ ECC)

§ Chip-to-chip networking — 5D Torus topology + external link — Each link 2 GB/s send + 2 GB/s receive — DMA, put/get, collecve operaons

4 LLNL-PRES-695482 Tradional BlueGene overall system integraon results in small footprint

4. Node card 3. Compute card 32 compute cards, 2. Module One single chip module, Optical modules, link chips, Single chip 16 GB DDR3 memory 1. Chip torus 16 cores

5b. I/O drawer 8 I/O cards 8 PCIe Gen2 slots 6. Rack 2 midplanes 1, 2, or 4 I/O drawers 7. System 20 PF/s

5a. Midplane 16 node cards

5 LLNL-PRES-695482 Sequoia and BlueGene/Q reflect lessons from previous BlueGene generaons § Communicaon locality through opmized MPI process placement is crical on 3D torus networks — Use of 5D torus reduces network diameter and reduces the importance of MPI process placement

§ Support for hardware opmized collecves should apply to subcommunicators as well as global operaons — Increased network communicaon contexts allows more applicaons to exploit hardware support for collecve operaons

§ Hardware support for network paroning minimizes jier

§ Mulple networks provide many benefits but also increase costs

These examples are network-centric; others reflect lessons throughout the system hardware and soware architecture

6 LLNL-PRES-695482 LLNL, with IBM, has developed Cardiod, a state- of-the-art cardiac electrophysiology simulaon § Mechanisms that lead to arrhythmia are not well understood; Contracon of heart is controlled by electrical behavior of heart cells

§ Mathemacal models reproduce component ionic currents of the acon potenal — System of non-linear equaon ODEs — TT06 includes 6 species and 14 gates — Negave currents depolarize (acvate) — Posive currents repolarize (return to rest)

§ Ability to run, at high resoluon, thousands instead of tens of heart beats enables detailed study of drug effects 2012 Gordon Bell finalist typifies the effort required to exploit capability fully

7 LLNL-PRES-695482 Cardoid achieves outstanding performance that enables nearly real-me heart beat simulaon § Measured peak performance: 11.84 PFlop/s (58.8% of peak) — 0.05 mm resoluon heart (3B ssue cells) — Ten million iteraons, dt = 4 usec — Performance of full simulaon loop, including I/O, measured with HPM

60 beats in 67.2 seconds 60 beats in 197.4 seconds

§ Extreme strong scaling limit: — 0.10 mm: 236 ssue cells/core. — 0.13 mm: 114 ssue cells/core

Opmized Cardioid is 50x faster than “naive” code

8 LLNL-PRES-695482 Cardiod represents a major advance in the state of the art of human heart simulaon

One minute of wall time 0.1 mm heart (370M tissue cells)

Cardioid Previous state of the art 18.2 seconds of simulation time

9 LLNL-PRES-695482 Cardoid achieves outstanding performance through detailed tuning to Sequoia’s architecture

§ Paroned cells over processes with an § Moved integer operaons to floang point upper bound on me (not on equal me) units to exploit SIMD units § Assigned diffusion work and reacon work to § No explicit network barrier different cores § L2 on node thread barriers § Transformed the potassium equaon to § Use low level SPI for halo data remove serializaon § Expensive 1D funcons in reacon model exchange between tasks (DMA) § Applicaon managed threads expressed with raonal approximates § § Single precision weights to reduce diffusion SIMDized diffusion stencil implementaon stencil use of L2 bandwidth § Zero flux boundary condions approximated § Hand unrolled to SIMDize loops over cells by method with no global solve § Sorted by cell type to improve SIMDizaon § High performance I/O is aware of § Sub-sorng of cells to increase sequenal/ BG/Q network topology vector load and storing of data. § Low overhead in-situ performance monitors § log funcon from libm replaced with custom § Assignment of threads to diffusion/reacon inlined funcons dependent on domain characteriscs § On the fly assembly of code to opmize data § Co-scheduled threads for improved dual movement at runme issue § Memory layout tuned to improve cache § Mulple diffusion implementaons to obtain performance opmal performance for various domains § Use of vector intrinsics and custom divides § Remote & local copies separated to improve bandwidth ulizaon

10 LLNL-PRES-695482 At largest scales, small soware overheads can significantly impact performance

370 Million Cells 1.6 Million Cores 1600 Flops/cell

60 us per iteration

MPI Halo Exchange SPI Halo Exchange OMP Fork/Join 60 us target L2 Atomic Barrier

0 50 100 150 200 250 300 Time (usec)

Direct use of message units and L2 atomic operaons minimizess overhead

11 LLNL-PRES-695482 The Sierra system that will replace Sequoia features a GPU-accelerated architecture Compute System Compute Rack 2.1 – 2.7 PB Memory Compute Node Standard 19” 120 -150 PFLOPS Warm water cooling 10 MW POWER® Architecture Processor NVIDIA®Volta™ NVMe-compatible PCIe 800GB SSD > 512 GB DDR4 + HBM Components Coherent Shared Memory IBM POWER • NVLink™

GPFS™ File System NVIDIA Volta Mellanox® Interconnect 120 PB usable storage • HBM Dual-rail EDR Infiniband® 1.2/1.0 TB/s R/W • NVLink bandwidth

12 LLNL-PRES-695482 Outstanding benchmark analysis by IBM and NVIDIA demonstrates the system’s usability CORAL APPLICATION PERFORMANCE PROJECTIONS

Scalable Science Benchmarks Throughput Benchmarks

14x 14x

12x 12x

10x 10x

8x 8x

6x 6x

4x 4x Performance Relative Relative Performance Relative

2x 2x

0x 0x CAM-SE UMT2013 AMG2013 MCB QBOX LSMS HACC NEKbone

CPU-only CPU + GPU CPU-only CPU + GPU

13

Figure 5: CORALProjecons included code changes that showed tractable benchmark projections show GPU-accelerated system is expected to deliver substantially higher performanceannotaon-based approach (i.e., at the system level compared to CPU-onlyOpenMP configuration) will be compeve .

13 TheLLNL-PRES-695482 demonstration of compelling, scalable performance at the system level across a wide range of applications proved to be one of the key factors in the U.S. DoE’s decision to build and Sierra on the GPU-accelerated OpenPOWER platform.

Conclusion Summit and Sierra are historic milestones in HPC’s efforts to reach exascale computing. With these new pre-exascale systems, the U.S. DoE maintains its leadership position, trailblazing the next generation of while allowing the nation to stay ahead in scientific discoveries and economic competitiveness.

The future of large-scale systems will inevitably be accelerated with throughput-oriented processors. Latency-optimized CPU-based systems have long hit a power wall that no longer delivers year-on-year performance increase. So while the question of “accelerator or not” is no longer in debate, other questions remain, such as CPU architecture, accelerator architecture, inter-node interconnect, intra- node interconnect, and heterogeneous versus self-hosted computing models.

With those questions in mind, the technological building blocks of these systems were carefully chosen with the focused goal of eventually deploying exascale supercomputers. The key building blocks that allow Summit and Sierra to meet this goal are:

• The Heterogeneous computing model • NVIDIA NVLink high-speed interconnect • NVIDIA GPU accelerator platform • IBM OpenPOWER platform

With the unveiling of the Summit and Sierra supercomputers, Oak Ridge National Laboratory and Lawrence Livermore National Laboratory have spoken loud and clear about the technologies that they believe will best carry the industry to exascale. 9

Sierra NRE will provide significant benefit to the final system

§ Center of Excellence § Open source compiler infrastructure § Motherboard design § System diagnostics § Water cooled compute nodes § System scheduling § HW resilience studies/ § investigation (NVIDIA) Burst buffer § GPFS performance and § Switch based collectives scalability § Hardware tag matching § Cluster management § GPU Direct and NVMe § Open source tools

14 LLNL-PRES-695482 Switch-based support for collecves further improves crical funconality

§ Reliable, scalable, general purpose primive,

applicable to mulple use cases SHArP Tree Root — In-network Tree based aggregaon mechanism — Large number of groups — Mulple simultaneous outstanding operaons

§ High performance collecve offload — Barrier, Reduce, All-Reduce — Sum, Min, Max, Min-loc, max-loc, OR, XOR, AND — Can overlap communicaon and computaon

§ Flexible mechanism reflects lessons learned SHArP Tree from BlueGene systems SHArP Tree Aggregation Node (Process running on HCA) SHArP Tree Endnode (Process running on HCA)

15 LLNL-PRES-695482 Inial results demonstrate that SHArP collecves improve performance significantly

§ OSU Allreduce 1PPN, 128 nodes

16 LLNL-PRES-695482 As seen with Cardoid, MPI soware overhead crically limits realized network performance § Realisc applicaons, parcularly in C++ oen use small messages — Realized message rate oen the key performance indicator — MPI provides lile ability to COALESCE these messages

§ MPI matching rules heavily impact realized message rate — Message envelops must match • Wild cards (MPI_ANY_SOURCE, MPI_ANY_TAG) increase envelop matching complexity and, thus, cost — Posted received must be matched in-order against the in-order posted sends Receiver Sender Tag=A, Communicator=B, Tag=A, Communicator=B, source=C, Time=X Destination=C, Time=Y

Tag=A, Communicator=B, Tag=A, Communicator=B, source=C, Time=X+D Destination=C, Time=Y+D’ Hardware message matching support can alleviate soware overhead

17 LLNL-PRES-695482 MPI tag matching operaons must appear to be performed atomically

§ Complexity/serialization of message matching limits the processing that can be performed on GPUs

18 LLNL-PRES-695482 Mellanox hardware will support efficient MPI tag matching § Offloaded to the ConnectX-5 HCA — Enables more of hardware bandwidth to translate to realized message rate — Full MPI tag matching as compute progresses — Rendezvous offload: large data delivery as compute progresses

§ Control can be passed between hardware and soware

§ Verbs tag matching support is being up-streamed

19 LLNL-PRES-695482 Deploying mulple networks exacerbates network hardware costs, which are already too high in large-scale systems § Sierra uses commodity cluster network soluon — Separate management (ethernet) and user (IB) networks — Single network for user traffic (point-to-point, collecves & file system traffic)

§ A single network for user traffic saves money but has other costs — Jier impact of other jobs’ file system traffic can be severe — Burst buffer strategy smooths file system bandwidth demand • File system traffic of a job now competes with its MPI traffic

§ Different types of network traffic are not equally crical — File system traffic “only” needs a guarantee of eventual compleon — Collecves oen crically limit overall performance — Other traffic classes also exist Quality-of-Service (QoS) mechanisms are necessary to achieve a network soluon that reduces network hardware costs while providing acceptable, consistent performance for all traffic classes

20 LLNL-PRES-695482 Preliminary results indicate that IB priority levels compensate for checkpoint traffic

Figure 21:Application (pF3D nearest neighbor exchange in Z, 16 processes per node) completion time for the various benchmarked configurations, in presence of checkpointing and in isolation.

Figure 21:Application (pF3D nearest neighbor exchange in Z, 16 processes per node) completion time for the various benchmarked configurations, in presence of checkpointing and in isolation.

Figure 22: Checkpointing rate in absence and in presence of application (pF3D nearest neighbor 21 LLNL-PRES-695482exchange in Z, 16 processes per node) for the various benchmarked configurations.

3.3 All-to-All Simulations on Fat Tree

FigureThe 22 all: Checkpointing to all communication rate in absence phases and ofin presence pF3D, dueof application to the mapping (pF3D nearest we haveneighbor chosen, exchangeinteract in with Z, 16 checkpointing processes per node) traffic for in the a varioussmaller benchmarked portion of the configurations. network. They do however exhibit similar behavior to that presented in the previous section for the nearest neighbor exchange3.3 All and- to the- All conclusions Simulations are the same. on In Fat the interestTree of brevity, we will only include the summary of the results. The figures present the following. TheFor allthe toall allto all communication exchange in the phases X directio of pF3D,n: due to the mapping we have chosen, interact withFigure checkpointing 23: pF3D completion traffic in timea smaller in the portion single processof the network. per node Theycase. do however exhibit similar behavior to that presented in the previous section for the nearest neighbor exchange Figure and the24 : conclusionscheckpointing are rate the in same.the pF3D In single the interest process of per brevity, node case we. will only include the summary of the results. The figures present the following. For the all to all exchange in the X direction: IBM Confidential Page 24 Figure 23: pF3D completion time in the single process per node case.

Figure 24: checkpointing rate in the pF3D single process per node case.

IBM Confidential Page 24

Sierra network hardware addresses lessons learned from previous LLNL systems § Flexibility of SHArP switch-based collecves will accelerate subcommunicators collecves and will allow jobs to share network § HCA MPI tag matching will reduce soware cost on crical path — Future systems should further accelerate message passing soware § QoS mechanisms are essenal with burst buffers or systems that shared network resources across jobs — Mulple networks might sll provide best soluon in some cases — Network paroning could sll be valuable on future systems § GPU Direct and NVMe reduce within node messaging impact § High capability nodes lead to a smaller network — Reduces importance of network paroning — LLNL CTS-1 with 2-to-1 tapered fat tree sll require careful task mapping Substanal research quesons sll remain for systems aer Sierra

22 LLNL-PRES-695482