2010 Annual Report
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Workload Management and Application Placement for the Cray Linux Environment™
TMTM Workload Management and Application Placement for the Cray Linux Environment™ S–2496–4101 © 2010–2012 Cray Inc. All Rights Reserved. This document or parts thereof may not be reproduced in any form unless permitted by contract or by written permission of Cray Inc. U.S. GOVERNMENT RESTRICTED RIGHTS NOTICE The Computer Software is delivered as "Commercial Computer Software" as defined in DFARS 48 CFR 252.227-7014. All Computer Software and Computer Software Documentation acquired by or for the U.S. Government is provided with Restricted Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7014, as applicable. Technical Data acquired by or for the U.S. Government, if any, is provided with Limited Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7013, as applicable. Cray and Sonexion are federally registered trademarks and Active Manager, Cascade, Cray Apprentice2, Cray Apprentice2 Desktop, Cray C++ Compiling System, Cray CX, Cray CX1, Cray CX1-iWS, Cray CX1-LC, Cray CX1000, Cray CX1000-C, Cray CX1000-G, Cray CX1000-S, Cray CX1000-SC, Cray CX1000-SM, Cray CX1000-HN, Cray Fortran Compiler, Cray Linux Environment, Cray SHMEM, Cray X1, Cray X1E, Cray X2, Cray XD1, Cray XE, Cray XEm, Cray XE5, Cray XE5m, Cray XE6, Cray XE6m, Cray XK6, Cray XK6m, Cray XMT, Cray XR1, Cray XT, Cray XTm, Cray XT3, Cray XT4, Cray XT5, Cray XT5h, Cray XT5m, Cray XT6, Cray XT6m, CrayDoc, CrayPort, CRInform, ECOphlex, LibSci, NodeKARE, RapidArray, The Way to Better Science, Threadstorm, uRiKA, UNICOS/lc, and YarcData are trademarks of Cray Inc. -
Cray XT and Cray XE Y Y System Overview
Crayyy XT and Cray XE System Overview Customer Documentation and Training Overview Topics • System Overview – Cabinets, Chassis, and Blades – Compute and Service Nodes – Components of a Node Opteron Processor SeaStar ASIC • Portals API Design Gemini ASIC • System Networks • Interconnection Topologies 10/18/2010 Cray Private 2 Cray XT System 10/18/2010 Cray Private 3 System Overview Y Z GigE X 10 GigE GigE SMW Fibre Channels RAID Subsystem Compute node Login node Network node Boot /Syslog/Database nodes 10/18/2010 Cray Private I/O and Metadata nodes 4 Cabinet – The cabinet contains three chassis, a blower for cooling, a power distribution unit (PDU), a control system (CRMS), and the compute and service blades (modules) – All components of the system are air cooled A blower in the bottom of the cabinet cools the blades within the cabinet • Other rack-mounted devices within the cabinet have their own internal fans for cooling – The PDU is located behind the blower in the back of the cabinet 10/18/2010 Cray Private 5 Liquid Cooled Cabinets Heat exchanger Heat exchanger (XT5-HE LC only) (LC cabinets only) 48Vdc flexible Cage 2 buses Cage 2 Cage 1 Cage 1 Cage VRMs Cage 0 Cage 0 backplane assembly Cage ID controller Interconnect 01234567 Heat exchanger network cable Cage inlet (LC cabinets only) connection air temp sensor Airflow Heat exchanger (slot 3 rail) conditioner 48Vdc shelf 3 (XT5-HE LC only) 48Vdc shelf 2 L1 controller 48Vdc shelf 1 Blower speed controller (VFD) Blooewer PDU line filter XDP temperature XDP interface & humidity sensor -
Pubtex Output 2011.12.12:1229
TM Using Cray Performance Analysis Tools S–2376–53 © 2006–2011 Cray Inc. All Rights Reserved. This document or parts thereof may not be reproduced in any form unless permitted by contract or by written permission of Cray Inc. U.S. GOVERNMENT RESTRICTED RIGHTS NOTICE The Computer Software is delivered as "Commercial Computer Software" as defined in DFARS 48 CFR 252.227-7014. All Computer Software and Computer Software Documentation acquired by or for the U.S. Government is provided with Restricted Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7014, as applicable. Technical Data acquired by or for the U.S. Government, if any, is provided with Limited Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7013, as applicable. Cray, LibSci, and PathScale are federally registered trademarks and Active Manager, Cray Apprentice2, Cray Apprentice2 Desktop, Cray C++ Compiling System, Cray CX, Cray CX1, Cray CX1-iWS, Cray CX1-LC, Cray CX1000, Cray CX1000-C, Cray CX1000-G, Cray CX1000-S, Cray CX1000-SC, Cray CX1000-SM, Cray CX1000-HN, Cray Fortran Compiler, Cray Linux Environment, Cray SHMEM, Cray X1, Cray X1E, Cray X2, Cray XD1, Cray XE, Cray XEm, Cray XE5, Cray XE5m, Cray XE6, Cray XE6m, Cray XK6, Cray XMT, Cray XR1, Cray XT, Cray XTm, Cray XT3, Cray XT4, Cray XT5, Cray XT5h, Cray XT5m, Cray XT6, Cray XT6m, CrayDoc, CrayPort, CRInform, ECOphlex, Gemini, Libsci, NodeKARE, RapidArray, SeaStar, SeaStar2, SeaStar2+, Sonexion, The Way to Better Science, Threadstorm, uRiKA, and UNICOS/lc are trademarks of Cray Inc. -
IUCAA CRAY CX1 Supercomputer Users Guide (V2.1)
IUCAA CRAY CX1 supercomputer Users Guide (v2.1) 1 Introduction Cray CX1 at the Inter-University Centre for Astronomy & Astrophysics (IUCAA) Pune is a desk- top/deskside supercomputer with 1 head node, 1 GPU node and 4 compute nodes. Each node has two 2.67 Ghz Intel Xeon(5650) processors, with every processor having six cores. In total there are 72 cores on the systems. Since there are twelve cores on every node, one can use twelve OpenMP hardware threads. However, the system supports more than 12 OpenMP threads on all the nodes (apart from the head node). Every node of the system has 24 GB RAM and in total the system has 144 GB RAM. At present there is around 5 TB disk space available for users on three dierent partitions, named /home, /data1/ and /data2. Keeping in mind that this is a small system and not more than ten users will be using it, every user can get 100 GB space in the home area. Data of common use (like WMAP and Planck data) is kept in a common area /data1/cmb_data and have the right permissions so that users can use it. The Cray CX1 is equipped with a 16 port RJ-45 10Base T, 100Base T, 1000Base T Gigabit Ethernet Switch, with the data transfer capability of 1 Gbps, supporting Ethernet, Fast Ethernet, and Gigabit Ethernet Data Link Protocol. The list of nodes is give in the Table (1). There are many software, packages and libraries already installed in the system and more (open 1 S. -
Use Style: Paper Title
Titan: Early experience with the Cray XK6 at Oak Ridge National Laboratory Arthur S. Bland, Jack C. Wells, Otis E. Messer, Oscar R. Hernandez, James H. Rogers National Center for Computational Sciences Oak Ridge National Laboratory Oak Ridge, Tennessee, 37831, USA [blandas, wellsjc, bronson, oscar, jrogers, at ORNL.GOV] Abstract— In 2011, Oak Ridge National Laboratory began an turbines, building a virtual nuclear reactor to increase power upgrade to Jaguar to convert it from a Cray XT5 to a Cray output and longevity of nuclear power plants, modeling the XK6 system named Titan. This is being accomplished in two ITER prototype fusion reactor, predicting impacts from phases. The first phase, completed in early 2012, replaced all of climate change, earthquakes and tsunamis, and improving the XT5 compute blades with XK6 compute blades, and the aerodynamics of vehicles to increase fuel economy. replaced the SeaStar interconnect with Cray’s new Gemini Access to Titan is available through several allocation network. Each compute node is configured with an AMD programs. More information about access is available Opteron™ 6274 16-core processors and 32 gigabytes of DDR3- through the OLCF web site at: 1600 SDRAM. The system aggregate includes 600 terabytes of http://www.olcf.ornl.gov/support/getting-started/. system memory. In addition, the first phase includes 960 The OLCF worked with Cray to design Titan to be an NVIDIA X2090 Tesla processors. In the second phase, ORNL will add NVIDIA’s next generation Tesla processors to exceptionally well-balanced system for modeling and increase the combined system peak performance to over 20 simulation at the highest end of high-performance PFLOPS. -
Cray HPCS Response 10/17/2013
Cray HPCS Response 10/17/2013 Cray Response to EEHPC Vendor Forum Slides presented on 12 September, 2013 Steven J. Martin Cray Inc. 1 10/17/2013 Copyright 2013 Cray Inc. Safe Harbor Statement This presentation may contain forward-looking statements that are based on our current expectations. Forward looking statements may include statements about our financial guidance and expected operating results, our opportunities and future potential, our product development and new product introduction plans, our ability to expand and penetrate our addressable markets and other statements that are not historical facts. These statements are only predictions and actual results may materially vary from those projected. Please refer to Cray's documents filed with the SEC from time to time concerning factors that could affect the Company and these forward-looking statements. 2 10/17/2013 Copyright 2013 Cray Inc. Legal Disclaimer Information in this document is provided in connection with Cray Inc. products. No license, express or implied, to any intellectual property rights is granted by this document. Cray Inc. may make changes to specifications and product descriptions at any time, without notice. All products, dates and figures specified are preliminary based on current expectations, and are subject to change without notice. Cray hardware and software products may contain design defects or errors known as errata, which may cause the product to deviate from published specifications. Current characterized errata are available on request. Cray uses codenames internally to identify products that are in development and not yet publically announced for release. Customers and other third parties are not authorized by Cray Inc. -
Leadership Computing Directions at Oak Ridge National Laboratory: Navigating the Transition to Heterogeneous Architectures
Leadership Computing Directions at Oak Ridge National Laboratory: Navigating the Transition to Heterogeneous Architectures International Symposium on Computing in Atmospheric Sciences September 8-12 , 2013 James J. Hack, Director National Center for Computational Sciences Oak Ridge National Laboratory Arthur S. “Buddy” Bland, OLCF Project Director Bronson Messer, OLCF Scientific Computing James Rogers, NCCS Director of Operations Jack Wells, NCCS Director of Science U.S. Department of Energy strategic priorities Incremental changes to existing technologies cannot meet these challenges – Transformational advances in energy technologies are needed – Transformational adaptation strategies will need to be implemented – Transformational changes to tools that allow a predictive explorations of paths forward Innovation Energy Security Investing in science, Providing clean, secure Safeguarding nuclear discovery and innovation energy and promoting and radiological materials, to provide solutions economic prosperity advancing responsible to pressing energy through energy efficiency legacy cleanup, and challenges and domestic maintaining nuclear forms of energy deterrence 2 Managed by UT-Battelle for the U.S. Department of Energy 3 Managed by UT-Battelle New for the York U.S. Department Magazine of Energy ExampleEnergy is of our the defining virtualization challenge challenge And a great example of success Big storm early next week (Oct 29) with wind and rain??? A number of computer models today were hinting at that, suggesting a tropical system now in the Caribbean may merge with a cold front in that time table, close to the East Coast. At the same time, 8 days is an eternity in storm forecasting, so plenty of time to watch this. Consider this an early “heads-up” that we’ll have an interesting an weather feature to monitor this week.. -
Leadership Computing at the National Center for Computational Science: Transitioning to Heterogeneous Architectures
Leadership Computing at the National Center for Computational Science: Transitioning to Heterogeneous Architectures 15th ECMWF Workshop on High Performance Computing in Meteorology 4 October 2012 James J. Hack, Director National Center for Computational Sciences Oak Ridge National Laboratory Arthur S. “Buddy” Bland, OLCF Project Director Bronson Messer, OLCF Scientific Computing James Rogers, NCCS Director of Operations Jack Wells, NCCS Director of Science U.S. Department of Energy strategic priorities Innovation Energy Security Investing in science, Providing clean, secure Safeguarding nuclear discovery and innovation energy and promoting and radiological materials, to provide solutions economic prosperity advancing responsible to pressing energy through energy efficiency legacy cleanup, and challenges and domestic maintaining nuclear forms of energy deterrence 2 Managed by UT-Battelle for the U.S. Department of Energy Energy is the defining challenge of our time The major driver for Global energy consumption – Climate change will increase 50% by 2030 – National security – Economic competitiveness – Quality of life Incremental changes to existing technologies cannot meet this challenge – Transformational advances in energy technologies are needed – Transformational adaptation strategies will need to be implemented – Transformational changes to tools enabling virtualization of strategies 33 Managed by UT-Battelle for the DepartmentU.S. Department of Energy of Energy ORNL has a long history in High Performance Computing 2007 IBM Blue -
Breakthrough Science Via Extreme Scalability
Breakthrough Science via Extreme Scalability Greg Clifford Segment Manager, Cray Inc. [email protected] • Cray’s focus • The requirement for highly scalable systems • Cray XE6 technology • The path to Exascale computing 10/6/2010 Oklahoma Supercomputing Symposium 2010 2 • 35 year legacy focused on building the worlds fastest computer. • 850 employees world wide • Growing in a tough economy • Cray XT6 first computer to deliver a PetaFLOP/s in a production environment (Jaguar system at Oakridge) • A full range of products • From the Cray CX1 to the Cray XE6 • Options includes: unsurpassed scalability, GPUs, SMP to 128 cores & 2 Tbytes, AMD and Intel, InfiniBand and Gemini, high performance IO, … 10/6/2010 Oklahoma Supercomputing Symposium 2010 3 Designed for “mission critical” HPC environments: “when you can not afford to be wrong” Sustainable performance on production applications Reliability Complete HPC environment. Focus on productivity Cray Linux Environment, Compilers, libraries, etc Partner with industry leaders (e.g. PGI, Platform, etc) Compatible with Open Source World Unsurpassed scalability/performance (compute, I/O and software) Proprietary system interconnect (Cray Gemini router) Performance on “grand challenge” applications 10/6/2010 Oklahoma Supercomputing Symposium 2010 4 Award(U. Tennessee/ORNL) Sep, 2007 Cray XT3: 7K cores, 40 TF Jun, 2008 Cray XT4: 18K cores,166 TF Aug 18, 2008 Cray XT5: 65K cores, 600 TF Feb 2, 2009 Cray XT5+: ~100K cores, 1 PF Oct, 2009 Kraken and Krakettes! NICS is specializing on true capability -
Sumit Gupta NVIDIA [email protected] Ian Miller Cray
Ian Miller Sumit Gupta Cray NVIDIA [email protected] [email protected] The Ease-of-Everything solution to ease the transition to HPC, increase engineering efficiency, and improve competitiveness Ease of configuration and purchase Ease of installation and deployment Ease of maintenance Pre-installed and tested combinations of industry-leading standard hardware, OS, and HPC management tools Upgradeable over time Complemented by Cray’s renowned quality of service and support No need for a dedicated computer room Compact deskside design Use of standard office power outlet (20A/110V or 16A/240V) Active noise reduction NR45 compliant Minimal power and cooling requirements Ability to mix and match compute, visualization, and storage blades according to a customer’s HPC needs Up to 8 blades per chassis (ability to combine up to 3 chassis) 2 sockets per blade (16 sockets per chassis) Up to 16 Intel Xeon 5500 quad-core processors per chassis Up to 64 cores per chassis 8 compute blades x 2 sockets x quad core Xeon processors = 64 cores Up to 48 GB of memory per blade, or 384 GB per chassis When 8 GB DIMMS are available, max of 96GB per blade and 768GB per chassis 1.7 TB of storage per storage blade, 6.8 TB per chassis Built-in Gigabit Ethernet Interconnect, optional InfiniBand GPU Computing Technology – Tesla with CUDA Up to 4 C1060 internal units per chassis Up to 4 S1070 external units per chassis CC55 (Dual Xeon) CV55-01 CT55-01 CS55-04 GPU Computing Compute Node Visualization Node Storage Node Node 1 Slot 2 Slots 2 Slots -
CUG 09-Program-042109-Fix NERSC.Indd
image courtesy of Sean Ahern (ORNL)•simulation byFredJaeger(ORNL) •radio-frequencyheating ofNSTXfusionplasma image courtesy ofSean cug 2009• atlanta •may 4-7 FUTURE THE COMPUTE Welcome Dear Friends, The National Center for Computational Sciences at Oak Ridge National Laboratory (ORNL) and the National Institute for Computational Sciences (NICS) are pleased to host CUG 2009. Much has changed since ORNL was the host in Knoxville for CUG 2004. The Cray XT line has really taken off, and parallelism has skyrocketed. Computational scientists and engineers are scrambling to take advantage of our potent new tools. But why? Why do we invest in these giant systems with many thousands of processors, cables, and disks? Why do we struggle to engineer the increasingly complex software required to harness their power? What do we expect to do with all this hardware, software, and electricity? Our theme for CUG 2009 is an answer to these questions... Compute the Future This is what many Cray users do. They design the next great superconductor, aircraft, and fusion-energy plant. They predict the efficiency of a biofuel, the track of an approaching hurricane, and the weather patterns that our grandchildren will encounter. They compute to know now, because they should not or must not wait. We welcome you to the OMNI Hotel in Atlanta for CUG 2009. This beautiful hotel and conference facility are in the world-famous CNN Center, overlooking Centennial Olympic Park, and within walking distance of a myriad of restaurants and attractions. We hope the convenience and excitement of Atlanta, the skills and attention of Cray experts, and the shared wisdom of your CUG colleagues will convince you. -
Cray Linux Environment™ (CLE) Software Release Overview Supplement
R Cray Linux Environment™ (CLE) Software Release Overview Supplement S–2497–5003 © 2010–2013 Cray Inc. All Rights Reserved. This document or parts thereof may not be reproduced in any form unless permitted by contract or by written permission of Cray Inc. U.S. GOVERNMENT RESTRICTED RIGHTS NOTICE The Computer Software is delivered as "Commercial Computer Software" as defined in DFARS 48 CFR 252.227-7014. All Computer Software and Computer Software Documentation acquired by or for the U.S. Government is provided with Restricted Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7014, as applicable. Technical Data acquired by or for the U.S. Government, if any, is provided with Limited Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7013, as applicable. Cray and Sonexion are federally registered trademarks and Active Manager, Cascade, Cray Apprentice2, Cray Apprentice2 Desktop, Cray C++ Compiling System, Cray CS300, Cray CX, Cray CX1, Cray CX1-iWS, Cray CX1-LC, Cray CX1000, Cray CX1000-C, Cray CX1000-G, Cray CX1000-S, Cray CX1000-SC, Cray CX1000-SM, Cray CX1000-HN, Cray Fortran Compiler, Cray Linux Environment, Cray SHMEM, Cray X1, Cray X1E, Cray X2, Cray XC30, Cray XD1, Cray XE, Cray XEm, Cray XE5, Cray XE5m, Cray XE6, Cray XE6m, Cray XK6, Cray XK6m, Cray XK7, Cray XMT, Cray XR1, Cray XT, Cray XTm, Cray XT3, Cray XT4, Cray XT5, Cray XT5h, Cray XT5m, Cray XT6, Cray XT6m, CrayDoc, CrayPort, CRInform, ECOphlex, LibSci, NodeKARE, RapidArray, The Way to Better Science, Threadstorm, uRiKA, UNICOS/lc, and YarcData are trademarks of Cray Inc.