HPC Case Study

Copyright 2012 FUJITSU LIMITED Customers of Large-scale HPC Systems

Customer Type No. of CPU Peak Perf. RIKEN (Kobe AICS) The 88,128 CPUs 11.28 PFlops Australian National University (NCI) x86 Cluster (CX400) 3,592 CPUs 1.2 PFlops System operation will start in early 2013 University of Tokyo FX10 4,800 CPUs 1.1 PFlops Kyushu University x86 Cluster (CX400), 510 TFlops 3,720 CPUs System operation will start in July 2012 FX10 182 TFlops HPC Wales, UK x86 Cluster > 2,000 CPUs > 300 TFlops Japan Atomic Energy Agency x86 Cluster, FX1, SMP > 4,568 CPUs 214 TFlops Institute for Molecular Science x86 Cluster (RX300), FX10 > 420 CPUs > 140 TFlops Japan Aerospace Exploration Agency FX1, SMP > 3,392 CPUs > 135 TFlops RIKEN (Wako Lab. RICC) x86 Cluster (RX200) > 2,048 CPUs 108 TFlops NAGOYA University x86 Cluster (HX600), FX1, SMP 1,504 CPUs 60 TFlops A*STAR, Singapore x86 Cluster (BX900) 900 CPUs > 45 TFlops A Manufacturer x86 Cluster > 2,600 CPUs > 77 TFlops B Manufacturer x86 Cluster > 2,000 CPUs > 38 TFlops Type definitions: FX10=PRIMEHPC FX10, x86 Cluster=Clusters based on PRIMERGY x86 server, SMP= SPARC Enterprise SMP server 1 Copyright 2012 FUJITSU LIMITED The University of Tokyo

 Key requirements  Increasing number of users and diverseness  Software compatibility with the K computer  1.4MW of power ceiling  System overview  PRIMEHPC FX10 (4,800 nodes (50 racks)) #18 on TOP500  Peak performance: 1.13 petaflops  Linpack performance: 1.04 petaflops (91.8% efficiency)  Focusing areas: earth science, astrophysics, seismology, weather modeling, materials science, energy, biology, hydrodynamics, solid-state physics…

2 Copyright 2012 FUJITSU LIMITED The University of Tokyo – System Overview

C計o算mノpuーtドe群 nod・インesタ,ラ Iクntテeィracブノtードive 群nodes Management servers PRIMEPHRPIMC EFXH10PC 5 0F筐体構X10 x成 50 racks (4,800 計算ノード + 300 IOノード) (4,800 compute nodes ) Job management Operation management [総理論演算性能 : 1.13PeakPFL POerPfoSr]mance: 1.13 petaflops Authentication [総主記憶容量 : 150TMBeymteo]ry capacity: 150 TB [インターコネクト : 6次I元nteメrッconnシュec/トーt: 6ラDス m] esh/torus - ”Tofu” PRIMERGY RX200S6 x 16 External connection router Local file system External Ethernet InfiniBand file system PRIMERGY RX300 S6 x 2 (MDS) network ETERNUS DX80 S2 x 150 (OST) network

Storage capacity: 1.1PB (RAID-5) Campus LAN Shared file system

PRIMERGY RX300 S6 x 8 (MDS) PRIMERGY RX300 S6 x 40 (OSS) End users ETERNUS DX80 S2 x 4 (MDT) Log-in nodes ETERNUS DX410 S2 x 80 (OST) InfiniBand Ethernet Storage capacity: 2.1PB (RAID-6) PRIMERGY RX300 S6 x 8 FibreChannel

3 Copyright 2012 FUJITSU LIMITED Kyushu University

Total peak performance: 691.7 teraflops  Features: Operations beginning: July 2012  Hybrid system of Fujitsu SPARC64 and x86 cluster  Software compatibility with the K computer  System  PRIMEHPC FX10 (768 nodes, SPARC64 IXfx)  Peak performance: 181.6 teraflops  High-performance Server System  PRIMERGY CX400 (2,952 CPUs, New Intel Xeon E5)  Peak Performance: 510.1 teraflops

4 Copyright 2012 FUJITSU LIMITED Kyushu University – System Overview

Local File System Supercomputer System High Performance ServerSystem (FEFS) PRIMEHPC FX10 (768 nodes) PRIMERGY CX400 (1,476 nodes) ETERNUS DX80 S2 181.6TFLOPS 510.1TFLOPS 345.6TB 24TB memory 184.5TB memory

InfiniBand Switch Gigabit Ethernet Switch

Shared File System (FEFS) LAN File Servers: PRIMERGY RX300 S7

4.0PB + 0.2PB Supercomputer: operation starts from July 2012 ETERNUS DX80 S2 High Performance Server System: from Sept 2012

5 Copyright 2012 FUJITSU LIMITED NCI-NF (Australia's national research computing service)  Key requirements  To improve the computational modeling capability in the research field below - Climate change - Ocean and marine - Earth system science - National water management research  Very high-energy efficiency, PUE is well under 1.20  System overview  PRIMERGY CX400 (Including CX250S1) : 3,592 nodes (50 racks) Peak performance : 1.2 Petaflops  PRIMEHPC FX10 : 96 nodes (1 rack)

6 Copyright 2010 FUJITSU LIMITED NCI-NF – System Overview

C計o算mノpuーtドe群 nod・インeタsラクティブノード群 Management servers PPRRIMIMEHEPRCG FYX C10X 54000筐体構 x 50成 racks (4(C,800X250S1 計算ノ ー: 3ド,592+ 300 node IOノsー, ド57) ,472 cores) Management, authentication servers: PRIMERGY RX300S7 x 7 Peak Performance: 1.2 Petaflops [総理論演算性能 : 1.13PFLOPS] Memory capacity: 150 TB 総主記憶容量 [ : 150InteTrBcyonnte]ect: Full bisection bandwidth (FDR) [インターコネクト : 6次元メッシュ/トーラス] External connection router Corroboration nodes PRIMEHPC FX10 Ethernet (96 compute nodes) InfiniBand network network Peak Performance: 22.7 Tflops Memory capacity: 3 TB Campus LAN Global file system

PRIMERGY RX300S7 x 6 (MDS) PRIMERGY RX300S7 x 30 (OSS) Login nodes End users DDN EF3015 x 3 (MDT) DDN SFA12000 x 5 (OST) Login, Data mover servers: InfiniBand PRIMERGY RX300S7 x 13 Ethernet Storage capacity: 12.6PB FibreChannel

7 Copyright 2010 FUJITSU LIMITED HPC Wales – A Grid of HPC Excellence

Motivation and background  Position Wales at the forefront of supercomputing  Promotion of research, technology and skills  Improvement of economic development  Creation of 400+ quality jobs, 10+ new business Implementation and rollout  Distributed HPC clusters among 15 academic sites  With central hubs, tier 1 and 2 sites  Portal for transparent, easy use of resources  Rollout completed by Q1 2012

8 Copyright 2012 FUJITSU LIMITED HPC Wales – Solution

Performance & Technology Solution Design  >1400 nodes PRIMERGY BX922S2  User-focused solution to access  Intel Xeon, X5650 and X5680 distributed HPC systems from desktop  Roadmap for upgrade browser  190 TFlops aggregated peak  Multiple components integrated into a performance consistent environment with a single sign-on  Infiniband, 10 / 1 Gb Ethernet, FCS  Data accessible across the entire  Eternus DX online SAN (home FS) infrastructure, automated movement  Parallel File System (up to 10 GB/s) driven by workflow  DDN Lustre  Collaborative sharing of information  Backup & Archiving and resources  Symantec, Quantum

9 Copyright 2012 FUJITSU LIMITED A*STAR

A*STAR PRIMERGY BX920 S2 at A*STAR  Singapore’s lead government agency  450 server blades (3888 cores)  Fostering world-class scientific research  45 Teraflops peak performance  Biomedical Sciences  91% of Linpack efficiency  Physical Sciences & Engineering  Spurs growth in key economic clusters

Fujitsu and A*STAR (IHPC)  R&D partnership to jointly develop  Applications  Technologies for the use of next-gen. supercomputer in  Computational Fluid Dynamics  Material Sciences

10 Copyright 2012 FUJITSU LIMITED Fujitsu HPC from workplace to #1 in TOP500

PRIMERGY x86 Clusters Celsius workstations PRIMEHPC FX10

11 Copyright 2012 FUJITSU LIMITED 12 Copyright 2012 FUJITSU LIMITED