Accelerating Computational Science Symposium 2012

Total Page:16

File Type:pdf, Size:1020Kb

Accelerating Computational Science Symposium 2012 Accelerating Computational Science Symposium 2012 Hosted by: The Oak Ridge Leadership Computing Facility at Oak Ridge National Laboratory The National Center for Supercomputing Applications at the University of Illinois at Urbana-Champaign The Swiss National Supercomputing Centre at ETH Zurich Scientists Assess a New Supercomputing Approach and Judge It a Win Graphics processing units speed up research codes on hybrid computing platforms Can scientists and engineers benefit from extreme-scale supercomputers that use application-code accelerators called GPUs (for graphics processing units)? Comparing GPU accelerators with today’s fastest central processing units, early results from diverse areas of research show 1.5- to 3-fold speedups for most codes, which in turn increase the realism of simulations and decrease time to results. With the availability of new, higher-performance GPUs later this year, such as the Kepler GPU chip to be installed in the 20-petaflop Titan supercomputer, application speedups are expected to be even more substantial. n innovation disrupts the The Accelerating Computational marketplace when it confers Science Symposium 2012 (ACSS Aan undeniable advantage 2012), held March 28–30 in Wash- over earlier technologies. Develop- ington, D.C., brought together about ing a disruptive innovation—such 100 experts in science, engineering, as the telephone that so long ago and computing from around the world replaced the telegraph—can be a to share transformational advances high-risk, high-rewards venture. enabled by hybrid supercomputers in Early in a product’s life, it may be diverse domains including chemistry, unclear whether customers will derive combustion, biology, nuclear fusion enough gain from a new technology and fission, and seismology. Partici- to justify its developers’ pain. Then pants also discussed the new breadth unprecedented success compels a and scope of research possible as the community to embrace the innovation. performance of petascale systems, With supercomputers, which must which execute quadrillions of calcula- balance performance and energy tions per second, continues to in- consumption, hybrid architectures crease. At the start of the symposium, incorporate high-performance, energy- neither the scientific community nor Jack Wells, ORNL efficient scientific-application-code the government officials charged with accelerators called graphics process- providing supercomputing facilities to sors to speed discoveries in fields ing units (GPUs) as well as traditional meet the needs of researchers knew from seismology to space and aid central processing units (CPUs). For whether hybrid computing architec- innovations such as next-generation scientists who must beat competitors tures conferred incontestable benefits catalysts, drugs, materials, engines, to discoveries in a publish-or-perish to the scientists and engineers using and reactors. The overwhelming world or engineers who must invent leadership-class computing systems consensus: GPUs accelerate a broad and improve products for the global to conduct digital experiments. range of computationally intensive ap- marketplace, the ability to more fully To understand the impact hybrid plications exploring the natural world, exploit massive parallelism is proving architectures are making on science, from subatomic particles to the vast advantageous enough for GPU/ attendees heard from two dozen cosmos, and the engineered world, CPU hybrids to displace CPU-only researchers who use a combination from turbines to advanced fuels. architectures. of traditional and accelerated proces- 1 “The symposium [was] motivated by society’s great need for advances in energy technologies and by the demonstrated achievements and tremendous potential for computa- tional science and engineering,” said meeting program chair Jack Wells, director of science at the Oak Ridge Leadership Computing Facility (OLCF), which co-hosted ACSS 2012 with the National Center for Supercomputing Applications (NCSA) and the Swiss National Supercomputing Centre (CSCS). Supercomputer vendor Cray Inc. and GPU inventor NVIDIA helped sponsor the event. “Compu- tational science on extreme-scale Monte Rosa, the Swiss National Supercomputing Centre’s 402-teraflop XE6 Cray hybrid computing architectures will supercomputer, uses only CPUs. The entire computational-science community benefits advance research and development in from efforts to optimize CPU performance. Image credit: Michele De Lorenzi/CSCS this decade, increase our understand- ing of the natural world, accelerate innovation, and as a result, increase economic opportunity,” Wells said. Concurring was James Hack, director of the National Center for Compu- tational Sciences, which houses the U.S. Department of Energy’s (DOE’s) Jaguar, the National Science Founda- tion’s (NSF’s) Kraken, and the National Oceanic and Atmospheric Administra- tion’s (NOAA’s) Gaea supercomputers. “As a community we’re really at a point where we need to embrace this disruptive approach to supercomput- ing,” he said. “This meeting was an Tödi, the Swiss National Supercomputing Centre’s 137-teraflop XK6 Cray opportunity to catalyze a broad dis- supercomputer, is the first GPU/CPU hybrid supercomputing system. cussion of how extreme-scale hybrid Image credit: Michele De Lorenzi/CSCS architectures are already accelerating progress in computationally focused system from Cray with Interlagos computational giant called Titan. So science research.” At its conclusion, CPUs and NVIDIA Fermi GPUs, has far the upgrade has converted the attendees had a better idea of where been in operation at CSCS since fall Cray XT5 system, an all-CPU model the community stood in exploiting 2011. Third, NCSA’s Blue Waters is a with an older, less-capable version of hybrid architectures. hybrid Cray system under construction the AMD processor, into a Cray XK6 at the University of Illinois at Urbana- system with Interlagos CPUs, twice The ACSS 2012 researchers Champaign that gives more weight to the memory of its predecessor, and presented results obtained with four CPUs than GPUs. It will have only the a more capable network. After the high-performance systems employ- Interlagos CPUs in 235 XE6 cabinets cabinets have been equipped with ing different computing strategies. and both CPUs and the NVIDIA Fermi NVIDIA’s newest Kepler GPUs by the First, CSCS’s 402-teraflop XE6 Cray GPUs in 30 XK6 cabinets. A testbed of end of 2012, the peak performance supercomputer, called Monte Rosa, at 48 XE6 cabinets has been operational will be at least 20 petaflops. In the the Swiss Federal Institute of Technol- since mid-March. Fourth, the OLCF’s meantime, ten of Jaguar’s cabinets ogy in Zurich (ETH Zurich) uses two 200-cabinet Jaguar supercomputer, have already gone hybrid to create a modern 16-core CPUs from AMD America’s fastest, is undergoing an testbed system called TitanDev. named Interlagos per compute node. upgrade at its Oak Ridge National Second, a two-cabinet XK6 super- Laboratory (ORNL) home that will computer called Tödi, the first hybrid transform it into a mostly hybrid 2 ALCC 2012 Workshop Report Compute Node Architectures of Select Cray Supercomputers XK6 Next XT5 XE6 (w/o XK6 -Gen GPU) XK AMD Istanbul CPU 2 (6 cores) AMD Interlagos CPU (16 cores) 2 1 1 1 NVIDIA Fermi 1 NVIDIA Kepler 1 Added Schulthess: “Given the a big avalanche coming that’s really The National Center for Supercomputing Applications’ Blue Waters, under applications that we’ve seen and the going to revolutionize computational construction at the University of Illinois benchmarks that were representative science. I think for these reasons we at Urbana-Champaign, is a hybrid of real science runs, I haven’t seen the really want to invest in hybrid comput- Cray system that gives more weight to case where the GPU doesn’t provide a ing from an application developer CPUs than GPUs. Image courtesy of performance increase. Really, it makes point of view.” the National Center for Supercomputing no sense not to have the GPU.” Applications and the Board of Trustees Work remains to determine if the of the University of Illinois The scientific results presented at the payoff for other applications and symposium were way ahead of what algorithms is great enough to develop At the start of ACSS 2012, managers the high-performance computing com- and port codes to hybrid, multicore of extreme-scale supercomputing munity had expected, Schulthess said. systems. “What we’re watching is a centers and program managers from Scientists from diverse fields of study bunch of pioneers, people who have DOE’s Office of Science wanted to were able to demonstrate impressive taken the first steps in embracing know if researchers would rather use results using hybrid architectures. and adopting what amounts to an an all-CPU XE6 or a hybrid GPU/ architectural sea change,” Hack said. CPU XK6. The results presented at “Compared to the past when we made the meeting laid that question to rest. major architectural changes, this one Most scientific application codes is different because the vendors are showed a 1.5- to 3-fold speedup on part of the solution. The partnering the GPU-equipped XK6 compared to that’s going on in terms of allowing the CPU-only XE6, with evidence of early adaptation or adoption of this a 6-fold speedup for one code. The new architecture includes working with greater computing speed makes pos- the vendors, something I hope will sible simulations of greater complexity continue. The kind of work that’s going and realism and quicker solutions to on now is identifying where the holes computationally demanding problems. are and where the opportunities are for “The Fermi [GPU from NVIDIA] delivers making progress.” higher performance than Interlagos [CPU from AMD],” said meeting NCSA Deputy Director Rob Penning- co-chair Thomas Schulthess, who is ton, a meeting co-chair, added that head of CSCS and a professor at ETH Thomas Schulthess, ETH Zurich and the partnership between domain and Zurich. “And with Fermi we’re talking CSCS computational scientists is also crucial about a GPU that is 1 or 2 years old “The XK architecture is looking like to providing whole solutions.
Recommended publications
  • Workload Management and Application Placement for the Cray Linux Environment™
    TMTM Workload Management and Application Placement for the Cray Linux Environment™ S–2496–4101 © 2010–2012 Cray Inc. All Rights Reserved. This document or parts thereof may not be reproduced in any form unless permitted by contract or by written permission of Cray Inc. U.S. GOVERNMENT RESTRICTED RIGHTS NOTICE The Computer Software is delivered as "Commercial Computer Software" as defined in DFARS 48 CFR 252.227-7014. All Computer Software and Computer Software Documentation acquired by or for the U.S. Government is provided with Restricted Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7014, as applicable. Technical Data acquired by or for the U.S. Government, if any, is provided with Limited Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7013, as applicable. Cray and Sonexion are federally registered trademarks and Active Manager, Cascade, Cray Apprentice2, Cray Apprentice2 Desktop, Cray C++ Compiling System, Cray CX, Cray CX1, Cray CX1-iWS, Cray CX1-LC, Cray CX1000, Cray CX1000-C, Cray CX1000-G, Cray CX1000-S, Cray CX1000-SC, Cray CX1000-SM, Cray CX1000-HN, Cray Fortran Compiler, Cray Linux Environment, Cray SHMEM, Cray X1, Cray X1E, Cray X2, Cray XD1, Cray XE, Cray XEm, Cray XE5, Cray XE5m, Cray XE6, Cray XE6m, Cray XK6, Cray XK6m, Cray XMT, Cray XR1, Cray XT, Cray XTm, Cray XT3, Cray XT4, Cray XT5, Cray XT5h, Cray XT5m, Cray XT6, Cray XT6m, CrayDoc, CrayPort, CRInform, ECOphlex, LibSci, NodeKARE, RapidArray, The Way to Better Science, Threadstorm, uRiKA, UNICOS/lc, and YarcData are trademarks of Cray Inc.
    [Show full text]
  • Pubtex Output 2011.12.12:1229
    TM Using Cray Performance Analysis Tools S–2376–53 © 2006–2011 Cray Inc. All Rights Reserved. This document or parts thereof may not be reproduced in any form unless permitted by contract or by written permission of Cray Inc. U.S. GOVERNMENT RESTRICTED RIGHTS NOTICE The Computer Software is delivered as "Commercial Computer Software" as defined in DFARS 48 CFR 252.227-7014. All Computer Software and Computer Software Documentation acquired by or for the U.S. Government is provided with Restricted Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7014, as applicable. Technical Data acquired by or for the U.S. Government, if any, is provided with Limited Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7013, as applicable. Cray, LibSci, and PathScale are federally registered trademarks and Active Manager, Cray Apprentice2, Cray Apprentice2 Desktop, Cray C++ Compiling System, Cray CX, Cray CX1, Cray CX1-iWS, Cray CX1-LC, Cray CX1000, Cray CX1000-C, Cray CX1000-G, Cray CX1000-S, Cray CX1000-SC, Cray CX1000-SM, Cray CX1000-HN, Cray Fortran Compiler, Cray Linux Environment, Cray SHMEM, Cray X1, Cray X1E, Cray X2, Cray XD1, Cray XE, Cray XEm, Cray XE5, Cray XE5m, Cray XE6, Cray XE6m, Cray XK6, Cray XMT, Cray XR1, Cray XT, Cray XTm, Cray XT3, Cray XT4, Cray XT5, Cray XT5h, Cray XT5m, Cray XT6, Cray XT6m, CrayDoc, CrayPort, CRInform, ECOphlex, Gemini, Libsci, NodeKARE, RapidArray, SeaStar, SeaStar2, SeaStar2+, Sonexion, The Way to Better Science, Threadstorm, uRiKA, and UNICOS/lc are trademarks of Cray Inc.
    [Show full text]
  • Use Style: Paper Title
    Titan: Early experience with the Cray XK6 at Oak Ridge National Laboratory Arthur S. Bland, Jack C. Wells, Otis E. Messer, Oscar R. Hernandez, James H. Rogers National Center for Computational Sciences Oak Ridge National Laboratory Oak Ridge, Tennessee, 37831, USA [blandas, wellsjc, bronson, oscar, jrogers, at ORNL.GOV] Abstract— In 2011, Oak Ridge National Laboratory began an turbines, building a virtual nuclear reactor to increase power upgrade to Jaguar to convert it from a Cray XT5 to a Cray output and longevity of nuclear power plants, modeling the XK6 system named Titan. This is being accomplished in two ITER prototype fusion reactor, predicting impacts from phases. The first phase, completed in early 2012, replaced all of climate change, earthquakes and tsunamis, and improving the XT5 compute blades with XK6 compute blades, and the aerodynamics of vehicles to increase fuel economy. replaced the SeaStar interconnect with Cray’s new Gemini Access to Titan is available through several allocation network. Each compute node is configured with an AMD programs. More information about access is available Opteron™ 6274 16-core processors and 32 gigabytes of DDR3- through the OLCF web site at: 1600 SDRAM. The system aggregate includes 600 terabytes of http://www.olcf.ornl.gov/support/getting-started/. system memory. In addition, the first phase includes 960 The OLCF worked with Cray to design Titan to be an NVIDIA X2090 Tesla processors. In the second phase, ORNL will add NVIDIA’s next generation Tesla processors to exceptionally well-balanced system for modeling and increase the combined system peak performance to over 20 simulation at the highest end of high-performance PFLOPS.
    [Show full text]
  • Titan Introduction and Timeline Bronson Messer
    Titan Introduction and Timeline Presented by: Bronson Messer Scientific Computing Group Oak Ridge Leadership Computing Facility Disclaimer: There is no contract with the vendor. This presentation is the OLCF’s current vision that is awaiting final approval. Subject to change based on contractual and budget constraints. We have increased system performance by 1,000 times since 2004 Hardware scaled from single-core Scaling applications and system software is the biggest through dual-core to quad-core and challenge dual-socket , 12-core SMP nodes • NNSA and DoD have funded much • DOE SciDAC and NSF PetaApps programs are funding of the basic system architecture scalable application work, advancing many apps research • DOE-SC and NSF have funded much of the library and • Cray XT based on Sandia Red Storm applied math as well as tools • IBM BG designed with Livermore • Computational Liaisons key to using deployed systems • Cray X1 designed in collaboration with DoD Cray XT5 Systems 12-core, dual-socket SMP Cray XT4 Cray XT4 Quad-Core 2335 TF and 1030 TF Cray XT3 Cray XT3 Dual-Core 263 TF Single-core Dual-Core 119 TF Cray X1 54 TF 3 TF 26 TF 2005 2006 2007 2008 2009 2 Disclaimer: No contract with vendor is in place Why has clock rate scaling ended? Power = Capacitance * Frequency * Voltage2 + Leakage • Traditionally, as Frequency increased, Voltage decreased, keeping the total power in a reasonable range • But we have run into a wall on voltage • As the voltage gets smaller, the difference between a “one” and “zero” gets smaller. Lower voltages mean more errors.
    [Show full text]
  • Introduction to the Oak Ridge Leadership Computing Facility for CSGF Fellows Bronson Messer
    Introduction to the Oak Ridge Leadership Computing Facility for CSGF Fellows Bronson Messer Acting Group Leader Scientific Computing Group National Center for Computational Sciences Theoretical Astrophysics Group Oak Ridge National Laboratory Department of Physics & Astronomy University of Tennessee Outline • The OLCF: history, organization, and what we do • The upgrade to Titan – Interlagos processors with GPUs – Gemini Interconnect – Software, etc. • The CSGF Director’s Discretionary Program • Questions and Discussion 2 ORNL has a long history in 2007 High Performance Computing IBM Blue Gene/P ORNL has had 20 systems 1996-2002 on the lists IBM Power 2/3/4 1992-1995 Intel Paragons 1985 Cray X-MP 1969 IBM 360/9 1954 2003-2005 ORACLE Cray X1/X1E 3 Today, we have the world’s most powerful computing facility Peak performance 2.33 PF/s #2 Memory 300 TB Disk bandwidth > 240 GB/s Square feet 5,000 Power 7 MW Dept. of Energy’s Jaguar most powerful computer Peak performance 1.03 PF/s #8 Memory 132 TB Disk bandwidth > 50 GB/s Square feet 2,300 National Science Kraken Power 3 MW Foundation’s most powerful computer Peak Performance 1.1 PF/s Memory 248 TB #32 Disk Bandwidth 104 GB/s Square feet 1,600 National Oceanic and Power 2.2 MW Atmospheric Administration’s NOAA Gaea most powerful computer 4 We have increased system performance by 1,000 times since 2004 Hardware scaled from single-core Scaling applications and system software is the biggest through dual-core to quad-core and challenge dual-socket , 12-core SMP nodes • NNSA and DoD have funded
    [Show full text]
  • Leadership Computing at the National Center for Computational Science: Transitioning to Heterogeneous Architectures
    Leadership Computing at the National Center for Computational Science: Transitioning to Heterogeneous Architectures 15th ECMWF Workshop on High Performance Computing in Meteorology 4 October 2012 James J. Hack, Director National Center for Computational Sciences Oak Ridge National Laboratory Arthur S. “Buddy” Bland, OLCF Project Director Bronson Messer, OLCF Scientific Computing James Rogers, NCCS Director of Operations Jack Wells, NCCS Director of Science U.S. Department of Energy strategic priorities Innovation Energy Security Investing in science, Providing clean, secure Safeguarding nuclear discovery and innovation energy and promoting and radiological materials, to provide solutions economic prosperity advancing responsible to pressing energy through energy efficiency legacy cleanup, and challenges and domestic maintaining nuclear forms of energy deterrence 2 Managed by UT-Battelle for the U.S. Department of Energy Energy is the defining challenge of our time The major driver for Global energy consumption – Climate change will increase 50% by 2030 – National security – Economic competitiveness – Quality of life Incremental changes to existing technologies cannot meet this challenge – Transformational advances in energy technologies are needed – Transformational adaptation strategies will need to be implemented – Transformational changes to tools enabling virtualization of strategies 33 Managed by UT-Battelle for the DepartmentU.S. Department of Energy of Energy ORNL has a long history in High Performance Computing 2007 IBM Blue
    [Show full text]
  • Warwick.Ac.Uk/Lib-Publications MENS T a T a G I MOLEM
    A Thesis Submitted for the Degree of PhD at the University of Warwick Permanent WRAP URL: http://wrap.warwick.ac.uk/102343/ Copyright and reuse: This thesis is made available online and is protected by original copyright. Please scroll down to view the document itself. Please refer to the repository record for this item for information to help you to cite it. Our policy information is available from the repository home page. For more information, please contact the WRAP Team at: [email protected] warwick.ac.uk/lib-publications MENS T A T A G I MOLEM U N IS IV S E EN RS IC ITAS WARW The Readying of Applications for Heterogeneous Computing by Andy Herdman A thesis submitted to The University of Warwick in partial fulfilment of the requirements for admission to the degree of Doctor of Philosophy Department of Computer Science The University of Warwick September 2017 Abstract High performance computing is approaching a potentially significant change in architectural design. With pressures on the cost and sheer amount of power, additional architectural features are emerging which require a re-think to the programming models deployed over the last two decades. Today's emerging high performance computing (HPC) systems are maximis- ing performance per unit of power consumed resulting in the constituent parts of the system to be made up of a range of different specialised building blocks, each with their own purpose. This heterogeneity is not just limited to the hardware components but also in the mechanisms that exploit the hardware components. These multiple levels of parallelism, instruction sets and memory hierarchies, result in truly heterogeneous computing in all aspects of the global system.
    [Show full text]
  • Cray Linux Environment™ (CLE) Software Release Overview Supplement
    R Cray Linux Environment™ (CLE) Software Release Overview Supplement S–2497–5003 © 2010–2013 Cray Inc. All Rights Reserved. This document or parts thereof may not be reproduced in any form unless permitted by contract or by written permission of Cray Inc. U.S. GOVERNMENT RESTRICTED RIGHTS NOTICE The Computer Software is delivered as "Commercial Computer Software" as defined in DFARS 48 CFR 252.227-7014. All Computer Software and Computer Software Documentation acquired by or for the U.S. Government is provided with Restricted Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7014, as applicable. Technical Data acquired by or for the U.S. Government, if any, is provided with Limited Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7013, as applicable. Cray and Sonexion are federally registered trademarks and Active Manager, Cascade, Cray Apprentice2, Cray Apprentice2 Desktop, Cray C++ Compiling System, Cray CS300, Cray CX, Cray CX1, Cray CX1-iWS, Cray CX1-LC, Cray CX1000, Cray CX1000-C, Cray CX1000-G, Cray CX1000-S, Cray CX1000-SC, Cray CX1000-SM, Cray CX1000-HN, Cray Fortran Compiler, Cray Linux Environment, Cray SHMEM, Cray X1, Cray X1E, Cray X2, Cray XC30, Cray XD1, Cray XE, Cray XEm, Cray XE5, Cray XE5m, Cray XE6, Cray XE6m, Cray XK6, Cray XK6m, Cray XK7, Cray XMT, Cray XR1, Cray XT, Cray XTm, Cray XT3, Cray XT4, Cray XT5, Cray XT5h, Cray XT5m, Cray XT6, Cray XT6m, CrayDoc, CrayPort, CRInform, ECOphlex, LibSci, NodeKARE, RapidArray, The Way to Better Science, Threadstorm, uRiKA, UNICOS/lc, and YarcData are trademarks of Cray Inc.
    [Show full text]
  • Overview of ORNL Leadership Computing Facility and Usage
    Overview of ORNL Leadership Computing Facility and Usage CScADS Summer Workshop Leadership Computing Platforms, Extreme-scale Applications, and Performance Strategies July 23, 2012 By Hai Ah Nam Scientific Computing Group Oak Ridge Leadership Computing Facility National Center for Computational Sciences Division Oak Ridge Leadership Computing Facility Mission The OLCF is a DOE Office of Science National User Facility whose mission is to enable breakthrough science by: • Fielding the most powerful capability computers for scientific research, • Building the required infrastructure to facilitate user access to these computers, • Selecting a few time-sensitive problems of national importance that can take advantage of these systems, • And partnering with these teams to deliver breakthrough science. 2 Breakthrough Science at Every Scale Nuclear Physics New Materials Nazarewicz et al., map the Lopez-Bezanilla et al., discover that boron-nitride nuclear driplines that mark monolayers are an ideal dielectric substrate for future the borders of nuclear nanoelectronic devices constructed with graphene as existence, predicting ~7000 the active layer bound nuclei, though only ~3000 have been observed. Nature 2012 Biochemistry Biofuels Ivanov et al., illuminate Smith et al., reveal how DNA replication the surface continues past a structure of lignin damaged site so a lesion clumps down to 1 can be repaired later angstrom Design Innovation Turbo Machinery Ramgen Power Efficiency Systems accelerates General Electric, for the their design of shock first
    [Show full text]
  • Cray Linux Environment™ (CLE) 4.0 Software Release Overview Supplement
    TM Cray Linux Environment™ (CLE) 4.0 Software Release Overview Supplement S–2497–4002 © 2010, 2011 Cray Inc. All Rights Reserved. This document or parts thereof may not be reproduced in any form unless permitted by contract or by written permission of Cray Inc. U.S. GOVERNMENT RESTRICTED RIGHTS NOTICE The Computer Software is delivered as "Commercial Computer Software" as defined in DFARS 48 CFR 252.227-7014. All Computer Software and Computer Software Documentation acquired by or for the U.S. Government is provided with Restricted Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7014, as applicable. Technical Data acquired by or for the U.S. Government, if any, is provided with Limited Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7013, as applicable. Cray, LibSci, and PathScale are federally registered trademarks and Active Manager, Cray Apprentice2, Cray Apprentice2 Desktop, Cray C++ Compiling System, Cray CX, Cray CX1, Cray CX1-iWS, Cray CX1-LC, Cray CX1000, Cray CX1000-C, Cray CX1000-G, Cray CX1000-S, Cray CX1000-SC, Cray CX1000-SM, Cray CX1000-HN, Cray Fortran Compiler, Cray Linux Environment, Cray SHMEM, Cray X1, Cray X1E, Cray X2, Cray XD1, Cray XE, Cray XEm, Cray XE5, Cray XE5m, Cray XE6, Cray XE6m, Cray XK6, Cray XMT, Cray XR1, Cray XT, Cray XTm, Cray XT3, Cray XT4, Cray XT5, Cray XT5h, Cray XT5m, Cray XT6, Cray XT6m, CrayDoc, CrayPort, CRInform, ECOphlex, Gemini, Libsci, NodeKARE, RapidArray, SeaStar, SeaStar2, SeaStar2+, Sonexion, The Way to Better Science, Threadstorm, uRiKA, and UNICOS/lc are trademarks of Cray Inc.
    [Show full text]
  • Early Experiences with the Cray XK6 Hybrid CPU and GPU MPP Platform
    Early experiences with the Cray XK6 hybrid CPU and GPU MPP platform Sadaf Alam, Jeffrey Poznanovic, Ugo Varetto and Nicola Bianchi (Swiss National Supercomputing Centre), Antonio Penya (UJI) and Nina Suvanphim (Cray Inc.) Cray User Group Meeting April 29 – May 3, 2012 Photo Gallery (more at: http://www.cscs.ch/) (Cray XK6, before March ‘12) (Cray XE6, before March ‘12) (New building, 2012) (New machine room) (Lake water cooling) Upgrades (processor, memory, interconnect) Cray XMT Maerhorn Meteoswiss Cray Cray XK6 XT4 and CrayXE6 Tödi (Magny-cours) 16-cores Opteron NVIDIA X2090 GPU Gemini Cray XE6 Palu 2x12-cores Opteron Gemini New CSCS Installaons Cray XT5 Cray XT5 Cray XE6 Rosa Rosa Rosa 2x4-cores 2x6-cores 2x16-cores Opteron Opteron Opterron Cray XT3 Cray XT3 SeaSrtarII Palu Palu SeaStarII Gemini 1-core 2-cores Opteron Opteron SeaStar SeaStar 2005 2006 2007 2008 2009 2010 2011 2012 CPU nodes: dual socket 6-cores AMD CPU: dual socket Intel Westmere Istanbul, 12-cores AMD Magny-cours GPU: NVIDI M2090 GPU devices: NVIDIA M2050, C2070, Interconnect: InfiniBand QDR (2 networks) S1070, GTX 480, GTX 285 Interconnect: Infiniband QDR Plaorms for programming, system soYware, management and operaonal R&D 5 High Performance High Productivity (HP2C) • BigDFT - Large scale Density Functional Electronic Structure Calculations in a Systematic Wavelet Basis Set; Prof. Stefan Goedecker, University of Basel • Cardiovascular - HPC for Cardiovascular System Simulations; Prof. Alfio Quarteroni, EPF Lausanne • COSMO-CCLM - Regional Climate and Weather Modeling on the Next Generations High- Performance Computers: Towards Cloud-Resolving Simulations; Dr. Isabelle Bey, ETH Zurich • Cosmology - Computational Cosmology on the Petascale; Prof.
    [Show full text]
  • AMD Opteron™ 6000 Series Platform Increases Reach, Demonstrates Virtualization Prowess, Drives Toward Core-Rich “Interlagos”
    2011-06-07 12:52 CEST AMD Opteron™ 6000 Series Platform Increases Reach, Demonstrates Virtualization Prowess, Drives Toward Core-Rich “Interlagos” SUNNYVALE, Calif. — June 07, 2011 — Since the launch of the AMD Opteron™ 6000 Series platform in 2010, AMD (NYSE: AMD) has continued to work with its OEM and channel partners to add to the platform’s portfolio. The AMD Opteron 6000 Series platform’s virtualization muscle has been noted in products such as the award-winning Dell PowerEdge R815 and the HP ProLiant DL585 G7. The AMD Opteron 6000 Series platform’s compatibility with the next generation 16-core processor, codenamed “Interlagos,” has gained the attention of customers in the early adopting world of high performance computing. “Our partners continue to offer great examples of how the AMD Opteron 6000 Series platform can deliver superior virtualization within an impressive power and thermal footprint,” said Patrick Patla, vice president and general manager, Server and Embedded Business, AMD. “The AMD Opteron 6000 Series platform offers budget-conscious IT managers performance and value today, and the promise of a future 16-core drop-in upgrade.” Partner Adoption AMD Opteron 6000 Series platform-based systems powered by AMD Opteron 6100 Series processors are available from OEMs and system builders including Acer, AMAX, Appro, Colfax, Cray Inc., Dawning, Dell, HP, IBM, Microway, NCS Technologies, Penguin, Rausch Netzwerktechnik, SGI, Silicon Mechanics, ZT Systems and more. In the past few months, new systems have been introduced by these
    [Show full text]