Exa Scale Computing: the Engine of Discovery

Total Page:16

File Type:pdf, Size:1020Kb

Exa Scale Computing: the Engine of Discovery EXA SCALE COMPUTING: THE ENGINE OF DISCOVERY SAI CHANDANA SIRISHA GORASA III B.Tech, Pragati Engineering College, ADB Road, Surampalem, East Godavari Dt. E-mail: [email protected] Abstract - This paper presents a detailed information about one of the levels of High Performance Computing namely Exascale computing. Computer architectures are expected to change to extend their support to Exa-scale computing in the near future. Exascale computing will have a profound impact on everyday life in the coming decades. It originated from petascale computing. It is simply considered as a transformation of petascale computing. One exaflop equals 1000 petaflops(1018flops per second)which is also known as a quintillion. Exascale computing is a milestone in the history of high performance computing. Keywords - High speed, Large scale, High Performance Computing, Floating point operations(Flops),Super Computers, Energy Consumption I. INTRODUCTION similar to the development of the human brain project.It could equate with the processing power of Computing is the act or the process of calculating the human brain. An Exascale computer can predict something using or connected with computers. climate change and thereby can withstand drought.It Exascale computing is the capability of computing could even be used to predict crime precisely and systems to perform at least one exaflops i.e., a billion with more accuracy than the current predictive billion calculations per second. Flop is the unit of policing systems. Drought-resistant plants and measurement of performance of computing. The first biofuels can be created. More sustainable future can petascale computer came into operation in the year be engineered with the help of stronger computers. 2008.Exascale super computer represents a vast Many problems can be overcome with ease with the increase in the performance of computing over the development of an Exascale computer. peta scale super computer. At a super computing Exascale computing is the term we use for the next conference in 2009, it is expected to be implemented 50- to 100-fold increase in speed over the fastest by 2018.Exascale super computer is assumed to be supercomputers which are in broad use today. An powerful enough to predict the future. Exascale super computer is a trillion times faster than the gigascale super computer.Achievement of such speeds is not an easy task.Several efforts are being carried out by different countries of the world to bring an exascale super computer into operation because of the benefits it brings to the performance of computing. III. DEVELOPMENT BY DIFFERENT COUNTRIES China: The two fastest super computers in the world Sunway TaihuLight and Tianhe-2(Milkyway-2) are owned by China and clock in at 93 and 33 petaflops. But the countries like US,Japan and other competing nations wanted to build the world’s first exascale super Figure 1: Different scales for measuring performance of computer.There is no computer in the world as computing powerful as the exascale super computer.According to the national plan for the next generation of high II. WHY EXA SCALE…??? performance computers, China will develop an exascale computer during the 13th Five-Year-Plan The more powerful and capable is the computer, the period (2016-2020).The exascale super computer is more realistic models it can create. Weather and planned to be named as Tianhe-3 by China. earthquakes are already predicted by the super Taiwan: computers we have. Complex biological systems The largest global center for the research and cannot be modelled with the current computing development of industrial and electronics technology power.Current super computers cannot equate the as well as the center of manufacturing for at least computing power of an Exascale super computer.It is 80% of all computer hardware technology in the Proceedings of IRAJ International Conference, 10th June, 2018, Hyderabad, India 17 Exa Scale Computing: The Engine of Discovery world is Taiwan.It initiated a lot of efforts by its the ARMv8 architecture with extensions it was co- various scientific organizations belonging to both designing with ARM Limited. government and private industries to build an Europe: exascale computer. Taiwan's National Center for European Union have already started building several High-Performance Computing took a first step projects which aimed at developing technologies and towards building the first Taiwanese exascale software for exascale computing in 2011.The projects supercomputer by funding construction of a new are the CRESTA project (Collaborative Research into intermediary supercomputer based on a full Exascale System ware, Tools and Applications), the technology transfer from Fujitsu corporation of Japan DEEP project (Dynamical Exascale Entry Platform), in June 2017..Taiwanese Foxconn and the project Mont-Blanc. A crucial and major Corporationrecently designed and built the largest European project which is based and concentrated on and fastest supercomputer in whole of Taiwan. This the transition of exascale is the MaX (Materials at the new Foxconn supercomputer is designed and built to Exascale) project.In 2015 the Scalable, Energy- serve as a wonderful invention in research and Efficient, Resilient and Transparent Software development towards the design and building of a Adaptation (SERT) project, a major research project state of the art Taiwanese exascale supercomputer. between the University of Manchester and the United States: Science and Technology Facilities Council (STFC). Because of the reason that building an exascale Daresbury Laboratory in Cheshire, was awarded c. computer is a national-level project,theU.S. £1million from the UK’s Engineering and Physical Department of Energy shelled out $258 million to six Sciences Research Council. The SERT project was different companies namely Hewlett-Packard, Intel, kept due to start in March 2015 andit will be funded Nvidia, Advanced Micro Devices, Cray, and IBM.All by Engineering and Physical Sciences Research these are currently working on the components that Council (EPSRC) under the Software for the Future II would one day go into building such a programme and the project will coordinate and system.Funding to the Institute for Advanced collaborate with the Numerical Analysis Group Architecturesto develop an exascale supercomputer (NAG), Cluster Vision and the STFC. was provided by the two United States of America India: governmental organizations within the US Government of India has proposed to commit 2.5 Department of Energy,the Office of Science and the billion USD in 2012for supercomputing research National Nuclear Security Administration in during the 12th five-year plan period (2012–2017). 2008.Sandia NationalLaboratory and the Oak Ridge The project will be handled by Indian Institute of National Laboratory also wanted to collaborate on Science (IISc), Bangalore.It was later known that exascale designs. Intel promised to develop an India attempts and is planning to develop a exascale technology by 2018.In order to fulfill its supercomputer with processing power in therange of promise,it purchased the InfiniBand product line from exaflops. Within the subsequent 5 years of approval,it QLogic for US $125million in January 2012.By 2012 will be handed over to C-DAC (Centre for the United States had allotted $126 million for Development of Advanced Computing) and will be exascale computing development.President Obama developed further. signed an executive order creating a National Strategic Computing Initiative calling for the The nation that developsan exascale super computer accelerated development of an exascale system and first will unlock all kinds of ways of predicting the funding research into post-semiconductor computing future and understanding the present. The country can on 29 July,2015. The Exascale Computing Project be far ahead of the rest of the world in terms of hopes to build an exascale computer by 2021. As part scientific and technological achievement, which in of the Department of Energy’s larger exascale turn translates to economic power. project, the NREL (National Renewable Energy Laboratory) located in Golden,Colorado is working to The following tables give information about the build predictive wind energy models that can work on fastest super computers developed as of November, an exascale-level machine by 2022. 2016 and take towards the exascale hardware: Japan: The RIKEN Advanced Institute for Computational Rank Name Country Type Science of Japan in the year 2013 began planning an Sunway 1 China Sunway26010 exascale system for 2020.Its main intension was to TaihuLight consume less than 30 megawatts. Later in Tianhe-2 2014,Fujitsu was awarded a contract by RIKEN.It Intel 2 (Milky China was to develop a next-generation supercomputer to Xeon/Xeon Phi succeed the K computer. In 2015, Fujitsu announced way -2) Opteron/Nvidia at the International Supercomputing Conference that 3 Titan USA this supercomputer will use processors implementing Kepler Proceedings of IRAJ International Conference, 10th June, 2018, Hyderabad, India 18 Exa Scale Computing: The Engine of Discovery IBM Power Reduction of Pollution: Exascale computing is 4 Sequoia USA BQC capable enough to reduce pollution caused by burning fossil fuels which is detrimental to our health. The Intel Xeon Phi 5 Cori USA environmental impact of burning fossil fuels is global warming. Carbon dioxide traps the heat in the Table 1: Fastest Super Computers as of November, 2016 atmosphere. It is emitted during fossil fuel burning. Approximately, 85 percent of the world’s energy is Peta Power Name Cores generated by burning fossil fuels. The only way to flops (MW) minimize pollution and optimize is by understanding Sunway and controlling the chemical process of combustion. 10,649,600 93 15 TaihuLight With the help of exascale computing, we expect it Tianhe-2 will be possible to increase the efficiency of (Milky 3,120,000 33.86 18 combustion systems in engines and gas turbines for way -2) transportation and power generation by potentially Titan 560,000 18 8 25-50%and to lower emissions.
Recommended publications
  • Materials Modelling and the Challenges of Petascale and Exascale
    Multiscale Materials Modelling on High Performance Computer Architectures Materials modelling and the challenges of petascale and exascale Andrew Emerson Cineca Supercomputing Centre, Bologna, Italy The project MMM@HPC is funded by the 7th Framework Programme of the European Commission within the Research Infrastructures 26/09/2013 with grant agreement number RI-261594. Contents Introduction to HPC HPC and the MMM@HPC project Petascale computing The Road to Exascale Observations 26/09/2013 A. Emerson, International Forum on Multiscale Materials Modelling, Bologna 2013 2 High Performance Computing High Performance Computing (HPC). What is it ? High-performance computing (HPC) is the use of parallel processing for running advanced application programs efficiently, reliably and quickly. The term applies especially to systems that function above a teraflop or 10 12 floating- point operations per second. (http://searchenterpriselinux.techtarget.com/definition/high -performance -computing ) A branch of computer science that concentrates on developing supercomputers and software to run on supercomputers. A main area of this discipline is developing parallel processing algorithms and software: programs that can be divided into little pieces so that each piece can be executed simultaneously by separate processors. (WEBOPEDIA ) 26/09/2013 A. Emerson, International Forum on Multiscale Materials Modelling, Bologna 2013 3 High Performance Computing Advances due to HPC, e.g. Molecular dynamics early 1990s . Lysozyme, 40k atoms 2006. Satellite tobacco mosaic virus (STMV). 1M atoms, 50ns 2008. Ribosome. 3.2M atoms, 230ns. 2011 . Chromatophore, 100M atoms (SC 2011) 26/09/2013 A. Emerson, International Forum on Multiscale Materials Modelling, Bologna 2013 4 High Performance Computing Cray-1 Supercomputer (1976) 80MHz , Vector processor → 250Mflops Cray XMP (1982) 2 CPUs+vectors, 400 MFlops “FERMI”, Bluegene/Q 168,000 cores 2.1 Pflops 26/09/2013 A.
    [Show full text]
  • Recent Developments in Supercomputing
    John von Neumann Institute for Computing Recent Developments in Supercomputing Th. Lippert published in NIC Symposium 2008, G. M¨unster, D. Wolf, M. Kremer (Editors), John von Neumann Institute for Computing, J¨ulich, NIC Series, Vol. 39, ISBN 978-3-9810843-5-1, pp. 1-8, 2008. c 2008 by John von Neumann Institute for Computing Permission to make digital or hard copies of portions of this work for personal or classroom use is granted provided that the copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise requires prior specific permission by the publisher mentioned above. http://www.fz-juelich.de/nic-series/volume39 Recent Developments in Supercomputing Thomas Lippert J¨ulich Supercomputing Centre, Forschungszentrum J¨ulich 52425 J¨ulich, Germany E-mail: [email protected] Status and recent developments in the field of supercomputing on the European and German level as well as at the Forschungszentrum J¨ulich are presented. Encouraged by the ESFRI committee, the European PRACE Initiative is going to create a world-leading European tier-0 supercomputer infrastructure. In Germany, the BMBF formed the Gauss Centre for Supercom- puting, the largest national association for supercomputing in Europe. The Gauss Centre is the German partner in PRACE. With its new Blue Gene/P system, the J¨ulich supercomputing centre has realized its vision of a dual system complex and is heading for Petaflop/s already in 2009. In the framework of the JuRoPA-Project, in cooperation with German and European industrial partners, the JSC will build a next generation general purpose system with very good price-performance ratio and energy efficiency.
    [Show full text]
  • Nascent Exascale Supercomputers Offer Promise, Present Challenges CORE CONCEPTS Adam Mann, Science Writer
    CORE CONCEPTS Nascent exascale supercomputers offer promise, present challenges CORE CONCEPTS Adam Mann, Science Writer Sometime next year, managers at the US Department Laboratory in NM. “We have to change our computing of Energy’s (DOE) Argonne National Laboratory in paradigms, how we write our programs, and how we Lemont, IL, will power up a calculating machine the arrange computation and data management.” size of 10 tennis courts and vault the country into That’s because supercomputers are complex a new age of computing. The $500-million main- beasts, consisting of cabinets containing hundreds of frame, called Aurora, could become the world’sfirst thousands of processors. For these processors to oper- “exascale” supercomputer, running an astounding ate as a single entity, a supercomputer needs to pass 1018, or 1 quintillion, operations per second. data back and forth between its various parts, running Aurora is expected to have more than twice the huge numbers of computations at the same time, all peak performance of the current supercomputer record while minimizing power consumption. Writing pro- holder, a machine named Fugaku at the RIKEN Center grams for such parallel computing is not easy, and the- for Computational Science in Kobe, Japan. Fugaku and orists will need to leverage new tools such as machine its calculation kin serve a vital function in modern learning and artificial intelligence to make scientific scientific advancement, performing simulations crucial breakthroughs. Given these challenges, researchers for discoveries in a wide range of fields. But the transition have been planning for exascale computing for more to exascale will not be easy.
    [Show full text]
  • Arxiv:2109.00082V1 [Cs.DC] 31 Aug 2021 Threshold of Exascale Computing
    Plan-based Job Scheduling for Supercomputers with Shared Burst Buffers Jan Kopanski and Krzysztof Rzadca Institute of Informatics, University of Warsaw Stefana Banacha 2, 02-097 Warsaw, Poland [email protected] [email protected] Preprint of the pa- Abstract. The ever-increasing gap between compute and I/O perfor- per accepted at the mance in HPC platforms, together with the development of novel NVMe 27th International storage devices (NVRAM), led to the emergence of the burst buffer European Conference concept—an intermediate persistent storage layer logically positioned on Parallel and Dis- between random-access main memory and a parallel file system. De- tributed Computing spite the development of real-world architectures as well as research (Euro-Par 2021), Lis- concepts, resource and job management systems, such as Slurm, provide bon, Portugal, 2021, only marginal support for scheduling jobs with burst buffer requirements, DOI: 10.1007/978-3- in particular ignoring burst buffers when backfilling. We investigate the 030-85665-6_8 impact of burst buffer reservations on the overall efficiency of online job scheduling for common algorithms: First-Come-First-Served (FCFS) and Shortest-Job-First (SJF) EASY-backfilling. We evaluate the algorithms in a detailed simulation with I/O side effects. Our results indicate that the lack of burst buffer reservations in backfilling may significantly deteriorate scheduling. We also show that these algorithms can be easily extended to support burst buffers. Finally, we propose a burst-buffer–aware plan-based scheduling algorithm with simulated annealing optimisation, which im- proves the mean waiting time by over 20% and mean bounded slowdown by 27% compared to the burst-buffer–aware SJF-EASY-backfilling.
    [Show full text]
  • The TOP500 List and Progress in High- Performance Computing
    COVER FEATURE GRAND CHALLENGES IN SCIENTIFIC COMPUTING The TOP500 List and Progress in High- Performance Computing Erich Strohmaier, Lawrence Berkeley National Laboratory Hans W. Meuer, University of Mannheim Jack Dongarra, University of Tennessee Horst D. Simon, Lawrence Berkeley National Laboratory For more than two decades, the TOP list has enjoyed incredible success as a metric for supercomputing performance and as a source of data for identifying technological trends. The project’s editors refl ect on its usefulness and limitations for guiding large-scale scientifi c computing into the exascale era. he TOP list (www.top.org) has served TOP500 ORIGINS as the de ning yardstick for supercomput- In the mid-s, coauthor Hans Meuer started a small ing performance since . Published twice a and focused annual conference that has since evolved year, it compiles the world’s largest instal- into the prestigious International Supercomputing Con- Tlations and some of their main characteristics. Systems ference (www.isc-hpc.com). During the conference’s are ranked according to their performance of the Lin- opening session, Meuer presented statistics about the pack benchmark, which solves a dense system of linear numbers, locations, and manufacturers of supercomput- equations. Over time, the data collected for the list has ers worldwide collected from vendors and colleagues in enabled the early identi cation and quanti cation of academia and industry. many important technological and architectural trends Initially, it was obvious that the supercomputer label related to high-performance computing (HPC).− should be reserved for vector processing systems from Here, we brie y describe the project’s origins, the companies such as Cray, CDC, Fujitsu, NEC, and Hitachi principles guiding data collection, and what has made that each claimed to have the fastest system for scienti c the list so successful during the two-decades-long tran- computation by some selective measure.
    [Show full text]
  • PETASCALE COMPUTING: Algorithms and Applications Edited by David A
    PETASCALE COMPUTING ALGORITHMS AND APPLICATIONS C9098_FM.indd 1 11/15/07 1:38:55 PM Chapman & Hall/CRC Computational Science Series SERIES EDITOR Horst Simon Associate Laboratory Director, Computing Sciences Lawrence Berkeley National Laboratory Berkeley, California, U.S.A. AIMS AND SCOPE This series aims to capture new developments and applications in the field of computational sci- ence through the publication of a broad range of textbooks, reference works, and handbooks. Books in this series will provide introductory as well as advanced material on mathematical, sta- tistical, and computational methods and techniques, and will present researchers with the latest theories and experimentation. The scope of the series includes, but is not limited to, titles in the areas of scientific computing, parallel and distributed computing, high performance computing, grid computing, cluster computing, heterogeneous computing, quantum computing, and their applications in scientific disciplines such as astrophysics, aeronautics, biology, chemistry, climate modeling, combustion, cosmology, earthquake prediction, imaging, materials, neuroscience, oil exploration, and weather forecasting. PUBLISHED TITLES PETASCALE COMPUTING: Algorithms and Applications Edited by David A. Bader C9098_FM.indd 2 11/15/07 1:38:55 PM PETASCALE COMPUTING ALGORITHMS AND APPLICATIONS EDITED BY DAVID A. BADER Georgia Institute of Technology Atlanta, U.S.A. C9098_FM.indd 3 11/15/07 1:38:56 PM Chapman & Hall/CRC Taylor & Francis Group 6000 Broken Sound Parkway NW, Suite 300 Boca Raton, FL 33487-2742 © 2008 by Taylor & Francis Group, LLC Chapman & Hall/CRC is an imprint of Taylor & Francis Group, an Informa business No claim to original U.S. Government works Printed in the United States of America on acid-free paper 10 9 8 7 6 5 4 3 2 1 International Standard Book Number-13: 978-1-58488-909-0 (Hardcover) This book contains information obtained from authentic and highly regarded sources.
    [Show full text]
  • SEVENTH FRAMEWORK PROGRAMME Research Infrastructures PRACE-2IP PRACE Second Implementation Project D5.1 Preliminary Guidance On
    SEVENTH FRAMEWORK PROGRAMME Research Infrastructures INFRA-2011-2.3.5 – Second Implementation Phase of the European High Performance Computing (HPC) service PRACE PRACE-2IP PRACE Second Implementation Project Grant Agreement Number: RI-283493 D5.1 Preliminary Guidance on Procurements and Infrastructure Final Version: 1.0 Author(s): Guillermo Aguirre, BSC Date: 21.02.2013 D5.1 Preliminary Guidance on Procurements and Infrastructure Project and Deliverable Information Sheet PRACE Project Project Ref. №: RI-283493 Project Title: PRACE Second Implementation Project Project Web Site: http://www.prace-project.eu Deliverable ID: < D5.1> Deliverable Nature: Report Deliverable Level: Contractual Date of Delivery: PU 28/02/2013 Actual Date of Delivery: 28/02/2013 EC Project Officer: Leonardo Flores Añover Document Control Sheet Title: Preliminary Guidance on Procurements and Infrastructure Document ID: D5.1 Version: <1.0 > Status: Final Available at: http://www.prace-project.eu Software Tool: Microsoft Word 2007 File(s): D5.1.docx Written by: Guillermo Aguirre, BSC Authorship Contributors: Francois Robin, CEA; Jean-Philippe Nominé, CEA; Ioannis Liabotis, GRNET; Norbert Meyer, PSNC; Radek Januszewski, PSNC; Andreas Johansson, SNIC-LIU; Eric Boyer, CINES; George Karagiannopoulos, GRNET; Marco Sbrighi, CINECA; Vladimir Slavnic, IPB; Gert Svensson, SNIC-KTH Reviewed by: Peter Stefan, NIIF Florian Berberich, PMO & FZJ Approved by: MB/TB Document Status Sheet Version Date Status Comments 0.1 16/01/2013 Draft First outline 0.2 22/01/2013 Draft Added contributions
    [Show full text]
  • Exascale Computing Study: Technology Challenges in Achieving Exascale Systems
    ExaScale Computing Study: Technology Challenges in Achieving Exascale Systems Peter Kogge, Editor & Study Lead Keren Bergman Shekhar Borkar Dan Campbell William Carlson William Dally Monty Denneau Paul Franzon William Harrod Kerry Hill Jon Hiller Sherman Karp Stephen Keckler Dean Klein Robert Lucas Mark Richards Al Scarpelli Steven Scott Allan Snavely Thomas Sterling R. Stanley Williams Katherine Yelick September 28, 2008 This work was sponsored by DARPA IPTO in the ExaScale Computing Study with Dr. William Harrod as Program Manager; AFRL contract number FA8650-07-C-7724. This report is published in the interest of scientific and technical information exchange and its publication does not constitute the Government’s approval or disapproval of its ideas or findings NOTICE Using Government drawings, specifications, or other data included in this document for any purpose other than Government procurement does not in any way obligate the U.S. Government. The fact that the Government formulated or supplied the drawings, specifications, or other data does not license the holder or any other person or corporation; or convey any rights or permission to manufacture, use, or sell any patented invention that may relate to them. APPROVED FOR PUBLIC RELEASE, DISTRIBUTION UNLIMITED. This page intentionally left blank. DISCLAIMER The following disclaimer was signed by all members of the Exascale Study Group (listed below): I agree that the material in this document reects the collective views, ideas, opinions and ¯ndings of the study participants only, and not those of any of the universities, corporations, or other institutions with which they are a±liated. Furthermore, the material in this document does not reect the o±cial views, ideas, opinions and/or ¯ndings of DARPA, the Department of Defense, or of the United States government.
    [Show full text]
  • SC20-Final-Program-V2.Pdf
    Table of Contents ACM Gordon Bell COVID Finalist Keynote ACM Gordon Bell Finalist More Than HPC Plenary ACM Student Research Competition: Panel Graduate Posters Paper ACM Student Research Competition: Research Posters Undergraduate Posters Scientific Visualization Awards Presentation & Data Analytics Showcase Birds of a Feather SCinet Booth Sessions State of the Practice Talk Doctoral Showcase Students@SC Early Career Program Test of Time Exhibitor Forum Tutorial Exhibits Virtual Student Cluster Competition Invited Talk Workshop Job Posting ACM Gordon Bell COVID Finalist (back to top) Thursday, November 19th 10:00 am - 12:00 pm Gordon Bell COVID-19 Prize Finalist Session 1 Session Description: Enabling Rapid COVID-19 Small Molecule Drug Design Through Scalable Deep Learning of Generative Models Sam Ade Jacobs (Lawrence Livermore National Laboratory), Tim Moon (Lawrence Livermore National Laboratory), Kevin McLoughlin (Lawrence Livermore National Laboratory), Derek Jones (Lawrence Livermore National Laboratory), David Hysom (Lawrence Livermore National Laboratory), Dong H. Ahn (Lawrence Livermore National Laboratory), John Gyllenhaal (Lawrence Livermore National Laboratory), Pythagoras Watson (Lawrence Livermore National Laboratory), Felice C. Lightsone (Lawrence Livermore National Laboratory), Jonathan E. Allen (Lawrence Livermore National Laboratory), Ian Karlin (Lawrence Livermore National Laboratory), Brian Van Essen (Lawrence Livermore National Laboratory) We improved the quality and reduced the time to produce machine-learned models for use in small molecule antiviral design. Our globally asynchronous multi-level parallel training approach strong scales to all of Sierra with up to 97.7% efficiency. We trained a novel, character-based Wasserstein autoencoder that produces a higher quality model trained on 1.613 billion compounds in 23 minutes while the previous state-of-the-art takes a day on 1 million compounds.
    [Show full text]
  • OLCF AR 2016-17 FINAL 9-7-17.Pdf
    Oak Ridge Leadership Computing Facility Annual Report 2016–2017 1 Outreach manager – Katie Bethea Writers – Eric Gedenk, Jonathan Hines, Katie Jones, and Rachel Harken Designer – Jason Smith Editor – Wendy Hames Photography – Jason Richards and Carlos Jones Stock images – iStockphoto™ Oak Ridge Leadership Computing Facility Oak Ridge National Laboratory P.O. Box 2008, Oak Ridge, TN 37831-6161 Phone: 865-241-6536 Email: [email protected] Website: https://www.olcf.ornl.gov Facebook: https://www.facebook.com/oakridgeleadershipcomputingfacility Twitter: @OLCFGOV The research detailed in this publication made use of the Oak Ridge Leadership Computing Facility, a US Department of Energy Office of Science User Facility located at DOE’s Oak Ridge National Laboratory. The Office of Science is the single largest supporter of basic research in the physical sciences in the United States and is working to address some of the most pressing challenges of our time. For more information, please visit science.energy.gov. 2 Contents LETTER In a Record 25th Year, We Celebrate the Past and Look to the Future 4 SCIENCE Streamlining Accelerated Computing for Industry 6 A Seismic Mapping Milestone 8 The Shape of Melting in Two Dimensions 10 A Supercomputing First for Predicting Magnetism in Real Nanoparticles 12 Researchers Flip Script for Lithium-Ion Electrolytes to Simulate Better Batteries 14 A Real CAM-Do Attitude 16 FEATURES Big Data Emphasis and New Partnerships Highlight the Path to Summit 18 OLCF Celebrates 25 Years of HPC Leadership 24 PEOPLE & PROGRAMS Groups within the OLCF 28 OLCF User Group and Executive Board 30 INCITE, ALCC, DD 31 SYSTEMS & SUPPORT Resource Overview 32 User Experience 34 Education, Outreach, and Training 35 ‘TitanWeek’ Recognizes Contributions of Nation’s Premier Supercomputer 36 Selected Publications 38 Acronyms 41 3 In a Record 25th Year, We Celebrate the Past and Look to the Future installed at the turn of the new millennium—to the founding of the Center for Computational Sciences at the US Department of Energy’s Oak Ridge National Laboratory.
    [Show full text]
  • At the Frontiers of Extreme Computing
    NOVEMBER 2011 SUPER- COMPUTERS AT THE FRONTIERS OF EXTREME COMPUTING PUBLISHED IN PARTNERSHIP WITH Research and Innovation with HPC Joint SMEs Laboratory HPC At the interface of computer science and mathematics, Inria researchers have spent 40 years establishing the scientific bases of a new field of knowledge: computational science. In inte- raction with other scientific disciplines, computational science offers new concepts, languages, methods and subjects for study that open new perspectives in the understanding of complex phenomena. High Performance Computing is a The work of this laboratory focuses Eventually, in order to boost techno- strategic topic for Inria, about thirty on development of algorithms and logy transfer from public research to Inria research teams are involved. software for computers at the peta- industry, which is part of Inria’s core flop scale and beyond. The laborato- mission, the institute has launched Inria has thus established large ry’s researchers carry out their work an «SME go HPC» Program, together scale strategic partnerships with- as part of the Blue Waters project. with GENCI, OSEO and four French Bull for the design of future HPC industry clusters (Aerospace Valley, architectures and with EDF R&D fo- It is also noteworthy that several Axelera, Minalogic, Systematic). cused on high performance simulation former Inria spin-off companies have The objective of the Program is to for energy applications. developed their business on this mar- bring high level expertise to SMEs wil- ket, such as Kerlabs, Caps Enterprise, ling to move to Simulation and HPC as At the international level, Inria and the Activeon or Sysfera.
    [Show full text]
  • Toward Loosely Coupled Programming on Petascale Systems
    Toward Loosely Coupled Programming on Petascale Systems Ioan Raicu*, Zhao Zhang+, Mike Wilde#+, Ian Foster#*+, Pete Beckman#, Kamil Iskra#, Ben Clifford+ *Department of Computer Science, University of Chicago, Chicago, IL, USA +Computation Institute, University of Chicago and Argonne National Laboratory, Chicago, IL, USA #Mathematics and Computer Science Division, Argonne National Laboratory, Argonne IL, USA [email protected], [email protected], {wilde,foster,beckman,iskra}@mcs.anl.gov, [email protected] Abstract— We have extended the Falkon lightweight task to-end application processes involving task coordination, execution framework to make loosely coupled programming on provenance tracking, and bookkeeping. Their approaches are petascale systems a practical and useful programming model. typically based on a model of loosely coupled computation, in This work studies and measures the performance factors which data is exchanged among tasks via files, databases or involved in applying this approach to enable the use of petascale XML documents, or a combination of these. Vast increases in systems by a broader user community, and with greater ease. data volume combined with the growing complexity of data Our work enables the execution of highly parallel computations analysis procedures and algorithms have rendered traditional composed of loosely coupled serial jobs with no modifications to manual processing and exploration unfavorable as compared the respective applications. This approach allows a new—and with modern high performance computing processes automated potentially far larger—class of applications to leverage petascale by scientific workflow systems. [3] systems, such as the IBM Blue Gene/P supercomputer. We present the challenges of I/O performance encountered in making The problem space can be partitioned into four main this model practical, and show results using both categories (Figure 1).
    [Show full text]