2017 HPC Annual Report Team Would Like to Acknowledge the Invaluable Assistance Provided by John Noe
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
The ASCI Red TOPS Supercomputer
The ASCI Red TOPS Supercomputer http://www.sandia.gov/ASCI/Red/RedFacts.htm The ASCI Red TOPS Supercomputer Introduction The ASCI Red TOPS Supercomputer is the first step in the ASCI Platforms Strategy, which is aimed at giving researchers the five-order-of-magnitude increase in computing performance over current technology that is required to support "full-physics," "full-system" simulation by early next century. This supercomputer, being installed at Sandia National Laboratories, is a massively parallel, MIMD computer. It is noteworthy for several reasons. It will be the world's first TOPS supercomputer. I/O, memory, compute nodes, and communication are scalable to an extreme degree. Standard parallel interfaces will make it relatively simple to port parallel applications to this system. The system uses two operating systems to make the computer both familiar to the user (UNIX) and non-intrusive for the scalable application (Cougar). And it makes use of Commercial Commodity Off The Shelf (CCOTS) technology to maintain affordability. Hardware The ASCI TOPS system is a distributed memory, MIMD, message-passing supercomputer. All aspects of this system architecture are scalable, including communication bandwidth, main memory, internal disk storage capacity, and I/O. Artist's Concept The TOPS Supercomputer is organized into four partitions: Compute, Service, System, and I/O. The Service Partition provides an integrated, scalable host that supports interactive users, application development, and system administration. The I/O Partition supports a scalable file system and network services. The System Partition supports system Reliability, Availability, and Serviceability (RAS) capabilities. Finally, the Compute Partition contains nodes optimized for floating point performance and is where parallel applications execute. -
An Overview of the Blue Gene/L System Software Organization
An Overview of the Blue Gene/L System Software Organization George Almasi´ , Ralph Bellofatto , Jose´ Brunheroto , Calin˘ Cas¸caval , Jose´ G. ¡ Castanos˜ , Luis Ceze , Paul Crumley , C. Christopher Erway , Joseph Gagliano , Derek Lieber , Xavier Martorell , Jose´ E. Moreira , Alda Sanomiya , and Karin ¡ Strauss ¢ IBM Thomas J. Watson Research Center Yorktown Heights, NY 10598-0218 £ gheorghe,ralphbel,brunhe,cascaval,castanos,pgc,erway, jgaglia,lieber,xavim,jmoreira,sanomiya ¤ @us.ibm.com ¥ Department of Computer Science University of Illinois at Urbana-Champaign Urabana, IL 61801 £ luisceze,kstrauss ¤ @uiuc.edu Abstract. The Blue Gene/L supercomputer will use system-on-a-chip integra- tion and a highly scalable cellular architecture. With 65,536 compute nodes, Blue Gene/L represents a new level of complexity for parallel system software, with specific challenges in the areas of scalability, maintenance and usability. In this paper we present our vision of a software architecture that faces up to these challenges, and the simulation framework that we have used for our experiments. 1 Introduction In November 2001 IBM announced a partnership with Lawrence Livermore National Laboratory to build the Blue Gene/L (BG/L) supercomputer, a 65,536-node machine de- signed around embedded PowerPC processors. Through the use of system-on-a-chip in- tegration [10], coupled with a highly scalable cellular architecture, Blue Gene/L will de- liver 180 or 360 Teraflops of peak computing power, depending on the utilization mode. Blue Gene/L represents a new level of scalability for parallel systems. Whereas existing large scale systems range in size from hundreds (ASCI White [2], Earth Simulator [4]) to a few thousands (Cplant [3], ASCI Red [1]) of compute nodes, Blue Gene/L makes a jump of almost two orders of magnitude. -
Advances in Ultrashort-Pulse Lasers • Modeling Dispersions of Biological and Chemical Agents • Centennial of E
October 2001 U.S. Department of Energy’s Lawrence Livermore National Laboratory Also in this issue: • More Advances in Ultrashort-Pulse Lasers • Modeling Dispersions of Biological and Chemical Agents • Centennial of E. O. Lawrence’s Birth About the Cover Computing systems leader Greg Tomaschke works at the console of the 680-gigaops Compaq TeraCluster2000 parallel supercomputer, one of the principal machines used to address large-scale scientific simulations at Livermore. The supercomputer is accessible to unclassified program researchers throughout the Laboratory, thanks to the Multiprogrammatic and Institutional Computing (M&IC) Initiative described in the article beginning on p. 4. M&IC makes supercomputers an institutional resource and helps scientists realize the potential of advanced, three-dimensional simulations. Cover design: Amy Henke About the Review Lawrence Livermore National Laboratory is operated by the University of California for the Department of Energy’s National Nuclear Security Administration. At Livermore, we focus science and technology on assuring our nation’s security. We also apply that expertise to solve other important national problems in energy, bioscience, and the environment. Science & Technology Review is published 10 times a year to communicate, to a broad audience, the Laboratory’s scientific and technological accomplishments in fulfilling its primary missions. The publication’s goal is to help readers understand these accomplishments and appreciate their value to the individual citizen, the nation, and the world. Please address any correspondence (including name and address changes) to S&TR, Mail Stop L-664, Lawrence Livermore National Laboratory, P.O. Box 808, Livermore, California 94551, or telephone (925) 423-3432. Our e-mail address is [email protected]. -
Safety and Security Challenge
SAFETY AND SECURITY CHALLENGE TOP SUPERCOMPUTERS IN THE WORLD - FEATURING TWO of DOE’S!! Summary: The U.S. Department of Energy (DOE) plays a very special role in In fields where scientists deal with issues from disaster relief to the keeping you safe. DOE has two supercomputers in the top ten supercomputers in electric grid, simulations provide real-time situational awareness to the whole world. Titan is the name of the supercomputer at the Oak Ridge inform decisions. DOE supercomputers have helped the Federal National Laboratory (ORNL) in Oak Ridge, Tennessee. Sequoia is the name of Bureau of Investigation find criminals, and the Department of the supercomputer at Lawrence Livermore National Laboratory (LLNL) in Defense assess terrorist threats. Currently, ORNL is building a Livermore, California. How do supercomputers keep us safe and what makes computing infrastructure to help the Centers for Medicare and them in the Top Ten in the world? Medicaid Services combat fraud. An important focus lab-wide is managing the tsunamis of data generated by supercomputers and facilities like ORNL’s Spallation Neutron Source. In terms of national security, ORNL plays an important role in national and global security due to its expertise in advanced materials, nuclear science, supercomputing and other scientific specialties. Discovery and innovation in these areas are essential for protecting US citizens and advancing national and global security priorities. Titan Supercomputer at Oak Ridge National Laboratory Background: ORNL is using computing to tackle national challenges such as safe nuclear energy systems and running simulations for lower costs for vehicle Lawrence Livermore's Sequoia ranked No. -
An Extensible Administration and Configuration Tool for Linux Clusters
An extensible administration and configuration tool for Linux clusters John D. Fogarty B.Sc A dissertation submitted to the University of Dublin, in partial fulfillment of the requirements for the degree of Master of Science in Computer Science 1999 Declaration I declare that the work described in this dissertation is, except where otherwise stated, entirely my own work and has not been submitted as an exercise for a degree at this or any other university. Signed: ___________________ John D. Fogarty 15th September, 1999 Permission to lend and/or copy I agree that Trinity College Library may lend or copy this dissertation upon request. Signed: ___________________ John D. Fogarty 15th September, 1999 ii Summary This project addresses the lack of system administration tools for Linux clusters. The goals of the project were to design and implement an extensible system that would facilitate the administration and configuration of a Linux cluster. Cluster systems are inherently scalable and therefore the cluster administration tool should also scale well to facilitate the addition of new nodes to the cluster. The tool allows the administration and configuration of the entire cluster from a single node. Administration of the cluster is simplified by way of command replication across one, some or all nodes. Configuration of the cluster is made possible through the use of a flexible, variables substitution scheme, which allows common configuration files to reflect differences between nodes. The system uses a GUI interface and is intuitively simple to use. Extensibility is incorporated into the system, by allowing the dynamic addition of new commands and output display types to the system. -
NNSA — Weapons Activities
Corporate Context for National Nuclear Security Administration (NS) Programs This section on Corporate Context that is included for the first time in the Department’s budget is provided to facilitate the integration of the FY 2003 budget and performance measures. The Department’s Strategic Plan published in September 2000 is no longer relevant since it does not reflect the priorities laid out in President Bush’s Management Agenda, the 2001 National Energy Policy, OMB’s R&D project investment criteria or the new policies that will be developed to address an ever evolving and challenging terrorism threat. The Department has initiated the development of a new Strategic Plan due for publication in September 2002, however, that process is just beginning. To maintain continuity of our approach that links program strategic performance goals and annual targets to higher level Departmental goals and Strategic Objectives, the Department has developed a revised set of Strategic Objectives in the structure of the September 2000 Strategic Plan. For more than 50 years, America’s national security has relied on the deterrent provided by nuclear weapons. Designed, built, and tested by the Department of Energy (DOE) and its predecessor agencies, these weapons helped win the Cold War, and they remain a key component of the Nation’s security posture. The Department’s National Nuclear Security Administration (NNSA) now faces a new and complex set of challenges to its national nuclear security missions in countering the threats of the 21st century. One of the most critical challenges is being met by the Stockpile Stewardship program, which is maintaining the effectiveness of our nuclear deterrent in the absence of underground nuclear testing. -
The Artisanal Nuke, 2014
The Artisanal Nuke Mary C. Dixon US Air Force Center for Unconventional Weapons Studies Maxwell Air Force Base, Alabama THE ARTISANAL NUKE by Mary C. Dixon USAF Center for Unconventional Weapons Studies 125 Chennault Circle Maxwell Air Force Base, Alabama 36112-6427 July 2014 Disclaimer The opinions, conclusions, and recommendations expressed or implied in this publication are those of the author and do not necessarily reflect the views of the Air University, Air Force, or Department of Defense. ii Table of Contents Chapter Page Disclaimer ................................................................................................... ii Table of Contents ....................................................................................... iii About the Author ......................................................................................... v Acknowledgements ..................................................................................... vi Abstract ....................................................................................................... ix 1 Introduction .............................................................................................. 1 2 Background ............................................................................................ 19 3 Stockpile Stewardship ........................................................................... 27 4 Opposition & Problems ......................................................................... 71 5 Milestones & Accomplishments .......................................................... -
Scalability and Performance of Salinas on the Computational Plant
Scalability and performance of salinas on the computa- tional plant Manoj Bhardwaj, Ron Brightwell & Garth Reese Sandia National Laboratories Abstract This paper presents performance results from a general-purpose, finite element struc- tural dynamics code called Salinas, on the Computational Plant (CplantTM), which is a large-scale Linux cluster. We compare these results to a traditional supercomputer, the ASCI/Red machine at Sandia National Labs. We describe the hardware and software environment of the Computational Plant, including its unique ability to support eas- ily switching a large section of compute nodes between multiple different independent cluster heads. We provide an overview of the Salinas application and present results from up to 1000 nodes on both machines. In particular, we discuss one of the chal- lenges related to scaling Salinas beyond several hundred processors on CplantTM and how this challenge was addressed and overcome. We have been able to demonstrate that the performance and scalability of Salinas is comparable to a proprietary large- scale parallel computing platform. 1 Introduction Parallel computing platforms composed of commodity personal computers (PCs) in- terconnected by gigabit network technology are a viable alternative to traditional pro- prietary supercomputing platforms. Small- and medium-sized clusters are now ubiqui- tous, and larger-scale procurements, such as those made recently by National Science Foundation for the Distributed Terascale Facility [1] and by Pacific Northwest National Lab [13], are becoming more prevalent. The cost effectiveness of these platforms has allowed for larger numbers of processors to be purchased. In spite of the continued increase in the number of processors, few real-world ap- plication results on large-scale clusters have been published. -
The Case for the Comprehensive Nuclear Test Ban Treaty
An Arms Control Association Briefing Book Now More Than Ever The Case for The Comprehensive nuClear TesT Ban TreaTy February 2010 Tom Z. Collina with Daryl G. Kimball An Arms Control Association Briefing Book Now More Than Ever The CAse for The Comprehensive nuCleAr TesT BAn Treaty February 2010 Tom Z. Collina with Daryl G. Kimball About the Authors Tom Z. Collina is Research Director at the Arms Control Association. He has over 20 years of professional experience in international security issues, previously serving as Director of the Global Security Program at the Union of Concerned Scientists and Executive Director of the Institute for Science and International Security. He was actively involved in national efforts to end U.S. nuclear testing in 1992 and international negotiations to conclude the CTBT in 1996. Daryl G. Kimball is Executive Director of the Arms Control Association. Previously he served as Executive Director of the Coalition to Reduce Nuclear Dangers, a consortium of 17 of the largest U.S. non-governmental organizations working together to strengthen national and international security by reducing the threats posed by nuclear weapons. He also worked as Director of Security Programs for Physicians for Social Responsibility, where he helped spearhead non-governmental efforts to win congressional approval for the 1992 nuclear test moratorium legislation, U.S. support for a “zero-yield” test ban treaty, and the U.N.’s 1996 endorsement of the CTBT. Acknowledgements The authors wish to thank our colleagues Pierce Corden, David Hafemeister, Katherine Magraw, and Benn Tannenbaum for sharing their expertise and reviewing draft text. -
Computer Architectures an Overview
Computer Architectures An Overview PDF generated using the open source mwlib toolkit. See http://code.pediapress.com/ for more information. PDF generated at: Sat, 25 Feb 2012 22:35:32 UTC Contents Articles Microarchitecture 1 x86 7 PowerPC 23 IBM POWER 33 MIPS architecture 39 SPARC 57 ARM architecture 65 DEC Alpha 80 AlphaStation 92 AlphaServer 95 Very long instruction word 103 Instruction-level parallelism 107 Explicitly parallel instruction computing 108 References Article Sources and Contributors 111 Image Sources, Licenses and Contributors 113 Article Licenses License 114 Microarchitecture 1 Microarchitecture In computer engineering, microarchitecture (sometimes abbreviated to µarch or uarch), also called computer organization, is the way a given instruction set architecture (ISA) is implemented on a processor. A given ISA may be implemented with different microarchitectures.[1] Implementations might vary due to different goals of a given design or due to shifts in technology.[2] Computer architecture is the combination of microarchitecture and instruction set design. Relation to instruction set architecture The ISA is roughly the same as the programming model of a processor as seen by an assembly language programmer or compiler writer. The ISA includes the execution model, processor registers, address and data formats among other things. The Intel Core microarchitecture microarchitecture includes the constituent parts of the processor and how these interconnect and interoperate to implement the ISA. The microarchitecture of a machine is usually represented as (more or less detailed) diagrams that describe the interconnections of the various microarchitectural elements of the machine, which may be everything from single gates and registers, to complete arithmetic logic units (ALU)s and even larger elements. -
Technical Issues in Keeping the Nuclear Stockpile Safe, Secure, and Reliable
Technical Issues in Keeping the Nuclear Stockpile Safe, Secure, and Reliable Marvin L. Adams Texas A&M University Sidney D. Drell Stanford University ABSTRACT The United States has maintained a safe, secure, and reliable nuclear stockpile for 16 years without relying on underground nuclear explosive tests. We argue that a key ingredient of this success so far has been the expertise of the scientists and engineers at the national labs with the responsibility for the nuclear arsenal and infrastructure. Furthermore for the foreseeable future this cadre will remain the critical asset in sustaining the nuclear enterprise on which the nation will rely, independent of the size of the stockpile or the details of its purpose, and to adapt to new requirements generated by new findings and changes in policies. Expert personnel are the foundation of a lasting and responsive deterrent. Thus far, the United States has maintained a safe, secure, and reliable nuclear stockpile by adhering closely to the designs of existing “legacy” weapons. It remains to be determined whether this is the best course to continue, as opposed to introducing new designs (as envisaged in the original RRW program), re-using previously designed and tested components, or maintaining an evolving combination of new, re-use, and legacy designs. We argue the need for a stewardship program that explores a broad spectrum of options, includes strong peer review in its assessment of options, and provides the necessary flexibility to adjust to different force postures depending on evolving strategic developments. U.S. decisions and actions about nuclear weapons can be expected to affect the nuclear policy choices of other nations – non-nuclear as well as nuclear – on whose cooperation we rely in efforts to reduce the global nuclear danger. -
W80-4 LRSO Missile Warhead
Description Accomplishments As part of the effort to keep the nuclear In 2015, the W80-4 LEP entered into Phase 6.2—feasibility study and down-select— arsenal safe, secure, and effective, the to establish program planning, mature technical requirements, and propose a full U.S. Air Force is developing a nuclear- suite of design options. In October 2017, the program entered Phase 6.2A—design capable Long-Range Standoff (LRSO) definition and cost study—to establish detailed cost estimates for all options and cruise missile. The LRSO is designed to a baseline from which the core cost estimate and detailed schedule are generated, replace the aging AGM-86 Air-Launched with appropriate risk mitigation. The W80-4 team is making significant progress: Cruise Missile, operational since 1982. The U.S. is extending the lifetime of its An effective team that includes the National Nuclear Security Administration warhead, the W80-1, through the W80-4 and the U.S. Air Force is in place; coordination is robust. Life Extension Program (LEP). The W80-1 will be refurbished and adapted to the new An extensive range of full-system tests, together with hundreds of small-scale tests, are in progress to provide confidence in the performance of W80-4 LEP LRSO missile as the W80-4. As the lead design options that include additively manufactured components. design agency for its nuclear explosive package, Lawrence Livermore, working LLNL leads DOE’s effort to develop and qualify the insensitive high explosive jointly with Sandia National Laboratories, formulated using newly produced constituents. Manufacturing of production- is developing options for the W80-4 that scale quantities of the new explosives is underway and on schedule.