BSC Roadmap for HPC in Europe
Total Page:16
File Type:pdf, Size:1020Kb

Load more
Recommended publications
-
An Operational Perspective on a Hybrid and Heterogeneous Cray XC50 System
An Operational Perspective on a Hybrid and Heterogeneous Cray XC50 System Sadaf Alam, Nicola Bianchi, Nicholas Cardo, Matteo Chesi, Miguel Gila, Stefano Gorini, Mark Klein, Colin McMurtrie, Marco Passerini, Carmelo Ponti, Fabio Verzelloni CSCS – Swiss National Supercomputing Centre Lugano, Switzerland Email: {sadaf.alam, nicola.bianchi, nicholas.cardo, matteo.chesi, miguel.gila, stefano.gorini, mark.klein, colin.mcmurtrie, marco.passerini, carmelo.ponti, fabio.verzelloni}@cscs.ch Abstract—The Swiss National Supercomputing Centre added depth provides the necessary space for full-sized PCI- (CSCS) upgraded its flagship system called Piz Daint in Q4 e daughter cards to be used in the compute nodes. The use 2016 in order to support a wider range of services. The of a standard PCI-e interface was done to provide additional upgraded system is a heterogeneous Cray XC50 and XC40 system with Nvidia GPU accelerated (Pascal) devices as well as choice and allow the systems to evolve over time[1]. multi-core nodes with diverse memory configurations. Despite Figure 1 clearly shows the visible increase in length of the state-of-the-art hardware and the design complexity, the 37cm between an XC40 compute module (front) and an system was built in a matter of weeks and was returned to XC50 compute module (rear). fully operational service for CSCS user communities in less than two months, while at the same time providing significant improvements in energy efficiency. This paper focuses on the innovative features of the Piz Daint system that not only resulted in an adaptive, scalable and stable platform but also offers a very high level of operational robustness for a complex ecosystem. -
A Scheduling Policy to Improve 10% of Communication Time in Parallel FFT
A scheduling policy to improve 10% of communication time in parallel FFT Samar A. Aseeri Anando Gopal Chatterjee Mahendra K. Verma David E. Keyes ECRC, KAUST CC, IIT Kanpur Dept. of Physics, IIT Kanpur ECRC, KAUST Thuwal, Saudi Arabia Kanpur, India Kanpur, India Thuwal, Saudi Arabia [email protected] [email protected] [email protected] [email protected] Abstract—The fast Fourier transform (FFT) has applications A. FFT Algorithm in almost every frequency related studies, e.g. in image and signal Forward Fourier transform is given by the following equa- processing, and radio astronomy. It is also used to solve partial differential equations used in fluid flows, density functional tion. 3 theory, many-body theory, and others. Three-dimensional N X (−ikxx) (−iky y) (−ikz z) 3 f^(k) = f(x; y; z)e e e (1) FFT has large time complexity O(N log2 N). Hence, parallel algorithms are made to compute such FFTs. Popular libraries kx;ky ;kz 3 perform slab division or pencil decomposition of N data. None The beauty of this sum is that k ; k ; and k can be summed of the existing libraries have achieved perfect inverse scaling x y z of time with (T −1 ≈ n) cores because FFT requires all-to-all independent of each other. These sums are done at different communication and clusters hitherto do not have physical all- to-all connections. With advances in switches and topologies, we now have Dragonfly topology, which physically connects various units in an all-to-all fashion. -
View Annual Report
Fellow Shareholders, Headlined by strong growth in revenue and profitability, we had one of our best years ever in 2015, executing across each of our major focus areas and positioning our company for continued growth into the future. We achieved another year of record revenue, growing by nearly 30 percent compared to 2014. In fact, our revenue in 2015 was more than three times higher than just four years prior — driven by growth in both our addressable market and market share. Over the last five years, we have transformed from a company solely focused on the high-end of the supercomputing market — where we are now the clear market leader — to a company with multiple product lines serving multiple markets. We provide our customers with powerful computing, storage and analytics solutions that give them the tools to advance their businesses in ways never before possible. During 2015, we installed supercomputing and storage solutions at a number of customers around the world. In the U.S., we completed the first phase of the massive new “Trinity” system at Los Alamos National Laboratory. This Cray XC40 supercomputer with Sonexion storage serves as the National Nuclear Security Administration’s flagship supercomputer, supporting all three of the NNSA’s national laboratories. We installed the first petaflop supercomputer in India at the Indian Institute of Science. In Europe, we installed numerous XC and storage solutions, including a significant expansion of the existing XC40 supercomputer at the University of Stuttgart in Germany. This new system nearly doubles their computing capacity and is currently the fastest supercomputer in Germany. -
TECHNICAL GUIDELINES for APPLICANTS to PRACE 17Th CALL
TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 17th CALL (T ier-0) The contributing sites and the corresponding computer systems for this call are: System Architecture Site (Country) Core Hours Minimum (node hours) request Joliot Curie - Bull Sequana X1000 GENCI@CEA 134 million 15 million core SKL (FR) (2.8 million) hours Joliot Curie - BULL Sequana GENCI@CEA 72 million 15 million core KNL X1000 (FR) (1,1 million) hours Hazel Hen Cray XC40 System GCS@HLRS 70 million 35 million core (DE) (2.9 million) hours JUWELS Multicore cluster GCS@JSC (DE) 70 million 35 million core (1.5 million) hours Marconi- Lenovo System CINECA (IT) 36 million 15 million core Broadwell (1 million) hours Marconi-KNL Lenovo System CINECA (IT) 612 million 30 million core (9 million) hours MareNostrum Lenovo System BSC (ES) 240 million 15 million core (5 million) hours Piz Daint Cray XC50 System CSCS (CH) 510 million 68 million core (7.5 million) hours Use of GPUs SuperMUC Lenovo NextScale/ GCS@LRZ (DE) 105 million 35 million core SuperMUC-NG Lenovo ThinkSystem (3.8 million) hours The site selection is done together with the specification of the requested computing time by the two sections at the beginning of the online form. The applicant can choose one or several machines as execution system, as long as proper benchmarks and resource request justification are provided on each of the requested systems. The parameters are listed in tables. The first column describes the field in the web online form to be filled in by the applicant. The remaining columns specify the range limits for each system. -
Cray XC40 Power Monitoring and Control for Knights Landing
Cray XC40 Power Monitoring and Control for Knights Landing Steven J. Martin, David Rush Matthew Kappel, Michael Sandstedt, Joshua Williams Cray Inc. Cray Inc. Chippewa Falls, WI USA St. Paul, MN USA {stevem,rushd}@cray.com {mkappel,msandste,jw}@cray.com Abstract—This paper details the Cray XC40 power monitor- This paper is organized as follows: In section II, we ing and control capabilities for Intel Knights Landing (KNL) detail blade-level Hardware Supervisory System (HSS) ar- based systems. The Cray XC40 hardware blade design for Intel chitecture changes introduced to enable improved telemetry KNL processors is the first in the XC family to incorporate enhancements directly related to power monitoring feedback gathering capabilities on Cray XC40 blades with Intel KNL driven by customers and the HPC community. This paper fo- processors. Section III provides information on updates to cuses on power monitoring and control directly related to Cray interfaces available for power monitoring and control that blades with Intel KNL processors and the interfaces available are new for Cray XC40 systems and the software released to users, system administrators, and workload managers to to support blades with Intel KNL processors. Section IV access power management features. shows monitoring and control examples. Keywords-Power monitoring; power capping; RAPL; energy efficiency; power measurement; Cray XC40; Intel Knights II. ENHANCED HSS BLADE-LEVEL MONITORING Landing The XC40 KNL blade incorporates a number of en- hancements over previous XC blade types that enable the I. INTRODUCTION collection of power telemetry in greater detail and with Cray has provided advanced power monitoring and control improved accuracy. -
Barcelona Supercomputing Center Puts Beauty of HPC on Display
CASE STUDY High Performance Computing (HPC) Beauty and High Performance Computing Under Glass at Barcelona Supercomputing Center Intel® Omni-Path Architecture and Intel® Xeon® Scalable Processor Family enable breakthrough European science using 13.7 petaFLOPS MareNostrum 4 Challenge MareNostrum 4 In publicly and privately funded computational science research, dollars (or • Spain’s 13.7 petaFLOPS Euros in this case) follow FLOPS. And when you’re one of the leading computing supercomputer contributes centers in Europe with a reputation around the world of highly reliable, leading- to the Partnership for edge technology resources, you look for the best in supercomputing in order to Advanced Computing in continue supporting breakthrough research. Thus, Barcelona Supercomputing Europe (PRACE) and supports Center (BSC), or Centro Nacional de Supercomputación, is driven to build leading the Spanish Supercomputing supercomputing clusters for its research clients in the public and private sectors. Network (RES) “We have the privilege of users coming back to us each year to run their projects,” said Sergi Girona, BSC’s Operations Department Director. “They return because we • 3,456 nodes of Intel® Xeon® reliably provide the technology and services they need year after year, and because Scalable Processors, plus our systems are of the highest level.” 0.5 petaFLOPS cluster of Intel® Xeon Phi™ Processors Supported by the Spanish and Catalan governments and funded by the Ministry (codenamed Knights Landing of Economy and Competitiveness with €34 million in 2015, BSC sought to take its MareNostrum 3 system to the next-generation of computing capabilities. It and Knights Hill) specified multiple clusters for both general computational needs of ongoing • Intel® Omni-Path Architecture research, and for development of next-generation codes for many integrated core interconnects multiple (MIC) supercomputing and tools for the Exascale computing era. -
Hpc in Europe
HPC IN EUROPE Organisation of public HPC resources Context • Focus on publicly-funded HPC resources provided primarily to enable scientific research and development at European universities and other publicly-funded research institutes • These resources are also intended to benefit industrial / commercial users by: • facilitating access to HPC • providing HPC training • sponsoring academic-industrial collaborative projects to exchange expertise and accelerate efficient commercial exploitation of HPC • Do not consider private sector HPC resources owned and managed internally by companies, e.g. in aerospace design & manufacturing, oil & gas exploration, fintech (financial technology), etc. European HPC Infrastructure • Structured provision of European HPC facilities: • Tier-0: European Centres (> petaflop machines) • Tier-1: National Centres • Tier-2: Regional/University Centres • Tiers planned as part of an EU Research Infrastructure Roadmap • This is coordinated through “PRACE” – http://prace-ri.eu PRACE Partnership foR Advanced Computing in Europe • International non-profit association (HQ office in Brussels) • Established in 2010 following ESFRI* roadmap to create a persistent pan-European Research Infrastructure (RI) of world-class supercomputers • Mission: enable high-impact scientific discovery and engineering research and development across all disciplines to enhance European competitiveness for the benefit of society. *European Strategy Forum on Reseach Infrastructures PRACE Partnership foR Advanced Computing in Europe Aims: • Provide access to leading-edge computing and data management resources and services for large-scale scientific and engineering applications at the highest performance level • Provide pan-European HPC education and training • Strengthen the European users of HPC in industry A Brief History of PRACE PRACE Phases & Objectives • Preparation and implementation of the PRACE RI was supported by a series of projects funded by the EU’s FP7 and Horizon 2020 funding programmes • 530 M€ of funding for the period 2010-2015. -
TOP500 Supercomputer Sites
7/24/2018 News | TOP500 Supercomputer Sites HOME | SEARCH | REGISTER RSS | MY ACCOUNT | EMBED RSS | SUPER RSS | Contact Us | News | TOP500 Supercomputer Sites http://top500.org/blog/category/feature-article/feeds/rss Are you the publisher? Claim or contact Browsing the Latest Browse All Articles (217 Live us about this channel Snapshot Articles) Browser Embed this Channel Description: content in your HTML TOP500 News Search Report adult content: 04/27/18--03:14: UK Commits a 0 0 Billion Pounds to AI Development click to rate The British government and the private sector are investing close to £1 billion Account: (login) pounds to boost the country’s artificial intelligence sector. The investment, which was announced on Thursday, is part of a wide-ranging strategy to make the UK a global leader in AI and big data. More Channels Under the investment, known as the “AI Sector Deal,” government, industry, and academia will contribute £603 million in new funding, adding to the £342 million already allocated in existing budgets. That brings the grand total to Showcase £945 million, or about $1.3 billion at the current exchange rate. The UK RSS Channel Showcase 1586818 government is also looking to increase R&D spending across all disciplines by 2.4 percent, while also raising the R&D tax credit from 11 to 12 percent. This is RSS Channel Showcase 2022206 part of a broader commitment to raise government spending in this area from RSS Channel Showcase 8083573 around £9.5 billion in 2016 to £12.5 billion in 2021. RSS Channel Showcase 1992889 The UK government policy paper that describes the sector deal meanders quite a bit, describing a lot of programs and initiatives that intersect with the AI investments, but are otherwise free-standing. -
Piz Daint”, One of the Most Powerful Supercomputers
The Swiss High-Performance Computing and Networking Thomas C. Schulthess | T. Schulthess !1 What is considered success in HPC? | T. Schulthess !2 High-Performance Computing Initiative (HPCN) in Switzerland High-risk & high-impact projects (www.hp2c.ch) Application driven co-design Phase III of pre-exascale supercomputing ecosystem Three pronged approach of the HPCN Initiative 2017 1. New, flexible, and efficient building 2. Efficient supercomputers 2016 Monte Rosa 3. Efficient applications Pascal based hybrid Cray XT5 2015 14’762 cores Upgrade to Phase II Cray XE6 K20X based hybrid 2014 Upgrade 47,200 cores Hex-core upgrade Phase I 22’128 cores 2013 Development & Aries network & multi-core procurement of 2012 petaflop/s scale supercomputer(s) 2011 2010 New 2009 building Begin construction complete of new building | T. Schulthess !3 FACT SHEET “Piz Daint”, one of the most powerful supercomputers A hardware upgrade in the final quarter of 2016 Thanks to the new hardware, researchers can run their simula- saw “Piz Daint”, Europe’s most powerful super- tions more realistically and more efficiently. In the future, big computer, more than triple its computing perfor- science experiments such as the Large Hadron Collider at CERN mance. ETH Zurich invested around CHF 40 million will also see their data analysis support provided by “Piz Daint”. in the upgrade, so that simulations, data analysis and visualisation can be performed more effi- ETH Zurich has invested CHF 40 million in the upgrade of “Piz ciently than ever before. Daint” – from a Cray XC30 to a Cray XC40/XC50. The upgrade in- volved replacing two types of compute nodes as well as the in- With a peak performance of seven petaflops, “Piz Daint” has tegration of a novel technology from Cray known as DataWarp. -
Accelerating Science with the NERSC Burst Buffer Early User Program
Accelerating Science with the NERSC Burst Buffer Early User Program Wahid Bhimji∗, Debbie Bard∗, Melissa Romanus∗y, David Paul∗, Andrey Ovsyannikov∗, Brian Friesen∗, Matt Bryson∗, Joaquin Correa∗, Glenn K. Lockwood∗, Vakho Tsulaia∗, Suren Byna∗, Steve Farrell∗, Doga Gursoyz, Chris Daley∗, Vince Beckner∗, Brian Van Straalen∗, David Trebotich∗, Craig Tull∗, Gunther Weber∗, Nicholas J. Wright∗, Katie Antypas∗, Prabhat∗ ∗ Lawrence Berkeley National Laboratory, Berkeley, CA 94720 USA, Email: [email protected] y Rutgers Discovery Informatics Institute, Rutgers University, Piscataway, NJ, USA zAdvanced Photon Source, Argonne National Laboratory, 9700 South Cass Avenue, Lemont, IL 60439, USA Abstract—NVRAM-based Burst Buffers are an important part their workflow, initial results and performance measurements. of the emerging HPC storage landscape. The National Energy We conclude with several important lessons learned from this Research Scientific Computing Center (NERSC) at Lawrence first application of Burst Buffers at scale for HPC. Berkeley National Laboratory recently installed one of the first Burst Buffer systems as part of its new Cori supercomputer, collaborating with Cray on the development of the DataWarp A. The I/O Hierarchy software. NERSC has a diverse user base comprised of over 6500 users in 700 different projects spanning a wide variety Recent hardware advancements in HPC systems have en- of scientific computing applications. The use-cases of the Burst abled scientific simulations and experimental workflows to Buffer at NERSC are therefore also considerable and diverse. We tackle larger problems than ever before. The increase in scale describe here performance measurements and lessons learned and complexity of the applications and scientific instruments from the Burst Buffer Early User Program at NERSC, which selected a number of research projects to gain early access to the has led to corresponding increase in data exchange, interaction, Burst Buffer and exercise its capability to enable new scientific and communication. -
Interconnect Your Future Enabling the Best Datacenter Return on Investment
Interconnect Your Future Enabling the Best Datacenter Return on Investment TOP500 Supercomputers, November 2015 InfiniBand FDR and EDR Continue Growth and Leadership The Most Used Interconnect On The TOP500 Supercomputing List EDR InfiniBand Now Used in 4 Systems on the List, with More EDR Systems Being Deployed by End of 2015 Will Result in More Than 2X the Number of EDR Systems Worldwide Compared to June 2015 . Mellanox Delivers Up To 99.8% System Efficiency with InfiniBand, 4X higher versus Ethernet • Delivering highest datacenter return on investment . For the first time many Chinese-based Ethernet Cloud/Web2.0 systems submitted and ranked on the list • Altering the trends perspective of high-performance computing systems on the TOP500 for Ethernet/InfiniBand connectivity © 2015 Mellanox Technologies 2 TOP500 Performance Trends 76% CAGR . Explosive high-performance computing market growth . Clusters continue to dominate with 85% of the TOP500 list Mellanox InfiniBand Solutions Provide the Highest Systems Utilization on the TOP500 © 2015 Mellanox Technologies 3 TOP500 Major Highlights . InfiniBand is the most used interconnect on the TOP500 list, with 237 systems, or 47.4% of the list . FDR InfiniBand is the most used technology on the TOP500 list, connecting 163 systems • 16% increase from Nov’14 to Nov’15 . EDR InfiniBand now used in 4 systems on the List, with more EDR systems being deployed by end of 2015 • More than 2X the number of EDR systems expected compared to July 2015 . InfiniBand enables the most efficient system on the list with 99.8% efficiency – record! • Delivering highest datacenter Return on Investment, 4X higher than 10GbE, 33% higher than Cray interconnect • InfiniBand enables the top 9 most efficient systems on the list, and 28 of the TOP30 . -
Cray System Monitoring: Successes, Requirements, and Priorities
Cray System Monitoring: Successes, Requirements, and Priorities x Ville Ahlgreny, Stefan Anderssonx, Jim Brandt , Nicholas P. Cardoz, Sudheer Chunduri∗, x xi Jeremy Enos∗∗, Parks Fieldsk, Ann Gentile , Richard Gerberyy, Joe Greenseid , Annette Greineryy, Bilel Hadri{, Yun (Helen) Heyy, Dennis Hoppex, Urpo Kailay, Kaki Kellyk, Mark Kleinz, x x Alex Kristiansen∗, Steve Leakyy, Mike Masonk, Kevin Pedretti , Jean-Guillaume Piccinaliz, Jason Repik , Jim Rogerszz, Susanna Salmineny, Mike Showerman∗∗, Cary Whitneyyy, and Jim Williamsk ∗Argonne Leadership Computing Facility (ALCF), Argonne, IL USA 60439 Email: (sudheerjalexk)@anl.gov yCenter for Scientific Computing (CSC - IT Center for Science), (02101 Espoo j 87100 Kajaani), Finland Email: (ville.ahlgrenjurpo.kailajsusanna.salminen)@csc.fi zSwiss National Supercomputing Centre (CSCS), 6900 Lugano, Switzerland Email: (cardojkleinjjgp)@cscs.ch xHigh Performance Computing Center Stuttgart (HLRS), 70569 Stuttgart, Germany Email: ([email protected],[email protected]) {King Abdullah University of Science and Technology (KAUST), Thuwal 23955, Saudia Arabia Email: [email protected] kLos Alamos National Laboratory (LANL), Los Alamos, NM USA 87545 Email: (parksjkakjmmasonjjimwilliams)@lanl.gov ∗∗National Center for Supercomputing Applications (NCSA), Urbana, IL USA 61801 Email: (jenosjmshow)@ncsa.illinois.edu yyNational Energy Research Scientific Computing Center (NERSC), Berkeley, CA USA 94720 Email: (ragerberjamgreinerjsleakjclwhitney)@lbl.gov zzOak Ridge National Laboratory (ORNL), Oak Ridge, TN USA 37830 Email: [email protected] x Sandia National Laboratories (SNL), Albuquerque, NM USA 87123 Email: (brandtjgentilejktpedrejjjrepik)@sandia.gov xi Cray Inc., Columbia, MD, USA 21045 Email: [email protected] Abstract—Effective HPC system operations and utilization in the procurement of a platform tailored to support the require unprecedented insight into system state, applications’ performance demands of a site’s workload.