The Bluegene/Q Supercomputer
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Professor Norman Christ — Curriculum Vitae Name: Norman H
Professor Norman Christ — Curriculum Vitae Name: Norman H. Christ Address: Department of Physics Columbia University New York, NY Telephone: 212 854 3307 Email address: [email protected] Current Position Professor of Physics Columbia University (1999-) Previous Positions Chair, Department of Physics Columbia University (1994-1999) Associate Professor Columbia University (1969-1974) Assistant Professor Columbia University (1966-1969) Instructor Princeton University (1967-1969) Education Ph.D., Columbia University (1966) B.A., Columbia University (summa cum laude) (1965) Honors and Awards: Salutatorian, Columbia College (1965) Sloan Fellowship (1967) American Physical Society Fellow (1981) Gordon Bell Prize (1998) Ken Wilson Lattice Award (2012) Publications: 1. N. H. Christ and T. D. Lee, Possible Tests of Cst and Tst Invariances in l± + N → l +Γ and A → B + e+e−, Phys. Rev. 143, 1310 (1966). 2. N. H. Christ and T. D. Lee, Receprocity Relations in Photopion Reac- tions, Phys. Rev. 148, 1520 (1966). 3. N. H. Christ and T. D. Lee, Possible CP Violation in K± → π±π0 + γ, Phys. Rev. 159, 1292 (1967). 4. N. H. Christ, A Renormalizable Theory of the Weak Interactions, Phys. Rev. 176, 2086, 176 (1968). 5. N. H. Christ, A Method for Computing Scattering Matrix Elements in a Class of Non-Renormalizable Field Theories, in Nonpolynomial Lagrangians, Renormalization and Gravity, Tracts in Mathematics and Natural Sciences, Vol 1, Gordon and Breach, 69 (1971). 6. N. H. Christ and T. D. Lee, CP Nonconservation and Inequalities Be- + − 0 0 tween µ µ and 2γ Decay Rates of KS and KL, Phys. Rev. D4, 209 (1971). 7. N. -
The Reliability Wall for Exascale Supercomputing Xuejun Yang, Member, IEEE, Zhiyuan Wang, Jingling Xue, Senior Member, IEEE, and Yun Zhou
. IEEE TRANSACTIONS ON COMPUTERS, VOL. *, NO. *, * * 1 The Reliability Wall for Exascale Supercomputing Xuejun Yang, Member, IEEE, Zhiyuan Wang, Jingling Xue, Senior Member, IEEE, and Yun Zhou Abstract—Reliability is a key challenge to be understood to turn the vision of exascale supercomputing into reality. Inevitably, large-scale supercomputing systems, especially those at the peta/exascale levels, must tolerate failures, by incorporating fault- tolerance mechanisms to improve their reliability and availability. As the benefits of fault-tolerance mechanisms rarely come without associated time and/or capital costs, reliability will limit the scalability of parallel applications. This paper introduces for the first time the concept of “Reliability Wall” to highlight the significance of achieving scalable performance in peta/exascale supercomputing with fault tolerance. We quantify the effects of reliability on scalability, by proposing a reliability speedup, defining quantitatively the reliability wall, giving an existence theorem for the reliability wall, and categorizing a given system according to the time overhead incurred by fault tolerance. We also generalize these results into a general reliability speedup/wall framework by considering not only speedup but also costup. We analyze and extrapolate the existence of the reliability wall using two representative supercomputers, Intrepid and ASCI White, both employing checkpointing for fault tolerance, and have also studied the general reliability wall using Intrepid. These case studies provide insights on how to mitigate reliability-wall effects in system design and through hardware/software optimizations in peta/exascale supercomputing. Index Terms—fault tolerance, exascale, performance metric, reliability speedup, reliability wall, checkpointing. ! 1INTRODUCTION a revolution in computing at a greatly accelerated pace [2]. -
LLNL Computation Directorate Annual Report (2014)
PRODUCTION TEAM LLNL Associate Director for Computation Dona L. Crawford Deputy Associate Directors James Brase, Trish Damkroger, John Grosh, and Michel McCoy Scientific Editors John Westlund and Ming Jiang Art Director Amy Henke Production Editor Deanna Willis Writers Andrea Baron, Rose Hansen, Caryn Meissner, Linda Null, Michelle Rubin, and Deanna Willis Proofreader Rose Hansen Photographer Lee Baker LLNL-TR-668095 3D Designer Prepared by LLNL under Contract DE-AC52-07NA27344. Ryan Chen This document was prepared as an account of work sponsored by an agency of the United States government. Neither the United States government nor Lawrence Livermore National Security, LLC, nor any of their employees makes any warranty, expressed or implied, or assumes any legal liability or responsibility for the accuracy, completeness, or usefulness of any information, apparatus, product, or process disclosed, or represents that its use would not infringe privately owned rights. Reference herein to any specific commercial product, process, or service by trade name, trademark, Print Production manufacturer, or otherwise does not necessarily constitute or imply its endorsement, recommendation, or favoring by the United States government or Lawrence Livermore National Security, LLC. The views and opinions of authors expressed herein do not necessarily state or reflect those of the Charlie Arteago, Jr., and Monarch Print Copy and Design Solutions United States government or Lawrence Livermore National Security, LLC, and shall not be used for advertising or product endorsement purposes. CONTENTS Message from the Associate Director . 2 An Award-Winning Organization . 4 CORAL Contract Awarded and Nonrecurring Engineering Begins . 6 Preparing Codes for a Technology Transition . 8 Flux: A Framework for Resource Management . -
Biomolecular Simulation Data Management In
BIOMOLECULAR SIMULATION DATA MANAGEMENT IN HETEROGENEOUS ENVIRONMENTS Julien Charles Victor Thibault A dissertation submitted to the faculty of The University of Utah in partial fulfillment of the requirements for the degree of Doctor of Philosophy Department of Biomedical Informatics The University of Utah December 2014 Copyright © Julien Charles Victor Thibault 2014 All Rights Reserved The University of Utah Graduate School STATEMENT OF DISSERTATION APPROVAL The dissertation of Julien Charles Victor Thibault has been approved by the following supervisory committee members: Julio Cesar Facelli , Chair 4/2/2014___ Date Approved Thomas E. Cheatham , Member 3/31/2014___ Date Approved Karen Eilbeck , Member 4/3/2014___ Date Approved Lewis J. Frey _ , Member 4/2/2014___ Date Approved Scott P. Narus , Member 4/4/2014___ Date Approved And by Wendy W. Chapman , Chair of the Department of Biomedical Informatics and by David B. Kieda, Dean of The Graduate School. ABSTRACT Over 40 years ago, the first computer simulation of a protein was reported: the atomic motions of a 58 amino acid protein were simulated for few picoseconds. With today’s supercomputers, simulations of large biomolecular systems with hundreds of thousands of atoms can reach biologically significant timescales. Through dynamics information biomolecular simulations can provide new insights into molecular structure and function to support the development of new drugs or therapies. While the recent advances in high-performance computing hardware and computational methods have enabled scientists to run longer simulations, they also created new challenges for data management. Investigators need to use local and national resources to run these simulations and store their output, which can reach terabytes of data on disk. -
Measuring Power Consumption on IBM Blue Gene/P
View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Springer - Publisher Connector Comput Sci Res Dev DOI 10.1007/s00450-011-0192-y SPECIAL ISSUE PAPER Measuring power consumption on IBM Blue Gene/P Michael Hennecke · Wolfgang Frings · Willi Homberg · Anke Zitz · Michael Knobloch · Hans Böttiger © The Author(s) 2011. This article is published with open access at Springerlink.com Abstract Energy efficiency is a key design principle of the Top10 supercomputers on the November 2010 Top500 list IBM Blue Gene series of supercomputers, and Blue Gene [1] alone (which coincidentally are also the 10 systems with systems have consistently gained top GFlops/Watt rankings an Rpeak of at least one PFlops) are consuming a total power on the Green500 list. The Blue Gene hardware and man- of 33.4 MW [2]. These levels of power consumption are al- agement software provide built-in features to monitor power ready a concern for today’s Petascale supercomputers (with consumption at all levels of the machine’s power distribu- operational expenses becoming comparable to the capital tion network. This paper presents the Blue Gene/P power expenses for procuring the machine), and addressing the measurement infrastructure and discusses the operational energy challenge clearly is one of the key issues when ap- aspects of using this infrastructure on Petascale machines. proaching Exascale. We also describe the integration of Blue Gene power moni- While the Flops/Watt metric is useful, its emphasis on toring capabilities into system-level tools like LLview, and LINPACK performance and thus computational load ne- highlight some results of analyzing the production workload glects the fact that the energy costs of memory references at Research Center Jülich (FZJ). -
Blue Gene®/Q Overview and Update
Blue Gene®/Q Overview and Update November 2011 Agenda Hardware Architecture George Chiu Packaging & Cooling Paul Coteus Software Architecture Robert Wisniewski Applications & Configurations Jim Sexton IBM System Technology Group Blue Gene/Q 32 Node Board Industrial Design BQC DD2.0 4-rack system 5D torus 3 © 2011 IBM Corporation IBM System Technology Group Top 10 reasons that you need Blue Gene/Q 1. Ultra-scalability for breakthrough science – System can scale to 256 racks and beyond (>262,144 nodes) – Cluster: typically a few racks (512-1024 nodes) or less. 2. Highest capability machine in the world (20-100PF+ peak) 3. Superior reliability: Run an application across the whole machine, low maintenance 4. Highest power efficiency, smallest footprint, lowest TCO (Total Cost of Ownership) 5. Low latency, high bandwidth inter-processor communication system 6. Low latency, high bandwidth memory system 7. Open source and standards-based programming environment – Red Hat Linux distribution on service, front end, and I/O nodes – Lightweight Compute Node Kernel (CNK) on compute nodes ensures scaling with no OS jitter, enables reproducible runtime results – Automatic SIMD (Single-Instruction Multiple-Data) FPU exploitation enabled by IBM XL (Fortran, C, C++) compilers – PAMI (Parallel Active Messaging Interface) runtime layer. Runs across IBM HPC platforms 8. Software architecture extends application reach – Generalized communication runtime layer allows flexibility of programming model – Familiar Linux execution environment with support for most -
TMA4280—Introduction to Supercomputing
Supercomputing TMA4280—Introduction to Supercomputing NTNU, IMF January 12. 2018 1 Outline Context: Challenges in Computational Science and Engineering Examples: Simulation of turbulent flows and other applications Goal and means: parallel performance improvement Overview of supercomputing systems Conclusion 2 Computational Science and Engineering (CSE) What is the motivation for Supercomputing? Solve complex problems fast and accurately: — efforts in modelling and simulation push sciences and engineering applications forward, — computational requirements drive the development of new hardware and software. 3 Computational Science and Engineering (CSE) Development of computational methods for scientific research and innovation in engineering and technology. Covers the entire spectrum of natural sciences, mathematics, informatics: — Scientific model (Physics, Biology, Medical, . ) — Mathematical model — Numerical model — Implementation — Visualization, Post-processing — Validation ! Feedback: virtuous circle Allows for larger and more realistic problems to be simulated, new theories to be experimented numerically. 4 Outcome in Industrial Applications Figure: 2004: “The Falcon 7X becomes the first aircraft in industry history to be entirely developed in a virtual environment, from design to manufacturing to maintenance.” Dassault Systèmes 5 Evolution of computational power Figure: Moore’s Law: exponential increase of number of transistors per chip, 1-year rate (1965), 2-year rate (1975). WikiMedia, CC-BY-SA-3.0 6 Evolution of computational power -
Introduction to the Practical Aspects of Programming for IBM Blue Gene/P
Introduction to the Practical Aspects of Programming for IBM Blue Gene/P Alexander Pozdneev Research Software Engineer, IBM June 22, 2015 — International Summer Supercomputing Academy IBM Science and Technology Center Established in 2006 • Research projects I HPC simulations I Security I Oil and gas applications • ESSL math library for IBM POWER • IBM z Systems mainframes I Firmware I Linux on z Systems I Software • IBM Rational software • Smarter Commerce for Retail • Smarter Cities 2 c 2015 IBM Corporation Outline 1 Blue Gene architecture in the HPC world 2 Blue Gene/P architecture 3 Conclusion 4 References 3 c 2015 IBM Corporation Outline 1 Blue Gene architecture in the HPC world Brief history of Blue Gene architecture Blue Gene in TOP500 Blue Gene in Graph500 Scientific applications Gordon Bell awards Blue Gene installation sites 2 Blue Gene/P architecture 3 Conclusion 4 References 4 c 2015 IBM Corporation Brief history of Blue Gene architecture • 1999 US$100M research initiative, novel massively parallel architecture, protein folding • 2003, Nov Blue Gene/L first prototype TOP500, #73 • 2004, Nov 16 Blue Gene/L racks at LLNL TOP500, #1 • 2007 Second generation Blue Gene/P • 2009 National Medal of Technology and Innovation • 2012 Third generation Blue Gene/Q • Features: I Low frequency, low power consumption I Fast interconnect balanced with CPU I Light-weight OS 5 c 2015 IBM Corporation Blue Gene in TOP500 Date # System Model Nodes Cores Jun’14 / Nov’14 3 Sequoia Q 96k 1.5M Jun’13 / Nov’13 3 Sequoia Q 96k 1.5M Nov’12 2 Sequoia -
Durham Unlocks Cosmological Secrets
Unlocking cosmological secrets Drawing on the power of high performance computing, Durham University and DiRAC scientists are expanding our understanding of the universe and its origins. Scientific Research | United Kingdom Business needs Solutions at a glance Durham University and the DiRAC HPC facility need • Dell EMC PowerEdge C6525 servers leading-edge high performance computing systems • AMD® EPYC™ 7H12 processors to power data- and compute-intensive cosmological • CoolIT® Systems Direct Liquid Cooling technology research. • Mellanox® HDR 200 interconnect Business results • Providing DiRAC researchers with world-class capabilities • Enabling a much greater level of discovery • Accelerating computational cosmology research • Keeping Durham University at the forefront cosmology research The COSMA8 cluster has The AMD EPYC processor used as many as in the COSMA8 cluster has cores 76,000 64 cores 64 per CPU CORES “This will primarily be used for testing, for getting code up Researching very big to scratch,” says Dr. Alastair Basden, technical manager for the COSMA high performance computing system at Durham questions University. “But then, coming very soon, we hope to place an order In scientific circles, more computing power can lead to bigger for another 600 nodes.” discoveries in less time. This is the case at Durham University, whose researchers are unlocking insights into our universe Deploying the full 600 nodes would provide a whopping 76,800 with powerful high performance computing clusters from Dell cores. “It’s a big uplift,” Dr. Basden notes. “COSMA7 currently is Technologies. about 12,000 cores. So COSMA8 will have six times as many cores. It will have more than twice as much DRAM as previous What is the universe? What is it made of? What is dark matter? systems, which means that we’ll be able to more than double What is dark energy? These are the types of questions explored the size of the simulations. -
The Blue Gene/Q Compute Chip
The Blue Gene/Q Compute Chip Ruud Haring / IBM BlueGene Team © 2011 IBM Corporation Acknowledgements The IBM Blue Gene/Q development teams are located in – Yorktown Heights NY, – Rochester MN, – Hopewell Jct NY, – Burlington VT, – Austin TX, – Bromont QC, – Toronto ON, – San Jose CA, – Boeblingen (FRG), – Haifa (Israel), –Hursley (UK). Columbia University . University of Edinburgh . The Blue Gene/Q project has been supported and partially funded by Argonne National Laboratory and the Lawrence Livermore National Laboratory on behalf of the United States Department of Energy, under Lawrence Livermore National Laboratory Subcontract No. B554331 2 03 Sep 2012 © 2012 IBM Corporation Blue Gene/Q . Blue Gene/Q is the third generation of IBM’s Blue Gene family of supercomputers – Blue Gene/L was announced in 2004 – Blue Gene/P was announced in 2007 . Blue Gene/Q – was announced in 2011 – is currently the fastest supercomputer in the world • June 2012 Top500: #1,3,7,8, … 15 machines in top100 (+ 101-103) – is currently the most power efficient supercomputer architecture in the world • June 2012 Green500: top 20 places – also shines at data-intensive computing • June 2012 Graph500: top 2 places -- by a 7x margin – is relatively easy to program -- for a massively parallel computer – and its FLOPS are actually usable •this is not a GPGPU design … – incorporates innovations that enhance multi-core / multi-threaded computing • in-memory atomic operations •1st commercial processor to support hardware transactional memory / speculative execution •… – is just a cool chip (and system) design • pushing state-of-the-art ASIC design where it counts • innovative power and cooling 3 03 Sep 2012 © 2012 IBM Corporation Blue Gene system objectives . -
Call for Applications for Dirac Community Development Director 1
DiRAC Health Data Science and AI Placement Opportunity DiRAC will awarD one Innovation Placement in 2021 in tHe area of HealtH Data Science anD tHe application of AI. The nominal length is 6 montHs anD Has to be completed by 30 September 2021. In tHis scheme a final year PhD stuDent or an early career researcHer can Have a funDed placement (up to £25k) witH the Getting It Right First Time (GIRFT) programme. GIRFT is funDeD by the UK Department of Health anD Social Care and is a collaboration between NHS England & NHS Improvement and the Royal National Orthopaedic Hospital NHS Trust. GIRFT uses comprehensive benchmarking data analysis to identify unwarranted variation in Healthcare provision and outcomes in National HealtH Service (NHS) Hospitals in EnglanD and combine this with deep dive visits to tHe Hospital by clinicians witH follow up on agreeD actions by an improvement team. The programme covers the majority of healthcare specialities. You have to be working on research that falls within the STFC remit in orDer to qualify for tHe placement; however, you can be funDeD by otHer organisations besides STFC, as long as tHe subject area is identifiable as being in Particle Physics, Astronomy & Cosmology, Solar Physics and Planetary Science, Astro-particle Physics, and Nuclear Physics. To check your eligibility please contact Jeremy Yates ([email protected]) anD Maria MarcHa ([email protected]). You must get your Supervisor or PIs permission before applying for this placement. It is allowed under UKRI’s rules, but only with your supervisor/PIs consent. -
NVIDIA Powers Next-Generation Supercomputer at University of Edinburgh
NVIDIA Powers Next-Generation Supercomputer at University of Edinburgh DiRAC Selects NVIDIA HDR InfiniBand Connected HGX Platform to Accelerate Scientific Discovery at Its Four Sites ISC—NVIDIA today announced that its NVIDIA HGX™ high performance computing platform will power Tursa, the new DiRAC supercomputer to be hosted by the University of Edinburgh. Optimized for computational particle physics, Tursa is the third of four DiRAC next-generation supercomputers formally announced that will be accelerated by one or more NVIDIA HGX platform technologies, including NVIDIA A100 Tensor Core GPUs, NVIDIA HDR 200Gb/s InfiniBand networking and NVIDIA Magnum IO™ software. The final DiRAC next-generation supercomputer is to feature NVIDIA InfiniBand networking. Tursa will allow researchers to carry out the ultra-high-precision calculations of the properties of subatomic particles needed to interpret data from massive particle physics experiments, such as the Large Hadron Collider. “DiRAC is helping researchers unlock the mysteries of the universe,” said Gilad Shainer, senior vice president of networking at NVIDIA. “Our collaboration with DiRAC will accelerate cutting-edge scientific exploration across a diverse range of workloads that take advantage of the unrivaled performance of NVIDIA GPUs, DPUs and InfiniBand in-network computing acceleration engines.” “Tursa is designed to tackle unique research challenges to unlock new possibilities for scientific modeling and simulation,” said Luigi Del Debbio, professor of theoretical physics at the University of Edinburgh and project lead for the DiRAC-3 deployment. “The NVIDIA accelerated computing platform enables the extreme-scaling service to propel new discoveries by precisely balancing network bandwidth and flops to achieve the unrivaled performance our research demands.” The Tursa supercomputer, built with Atos and expected to go into operation later this year, will feature 448 NVIDIA A100 Tensor Core GPUs and include 4x NVIDIA HDR 200Gb/s InfiniBand networking adapters per node.