Performance Analysis of Sweep3d on Blue Gene/P with Scalasca
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Survey of Computer Architecture
Architecture-aware Algorithms and Software for Peta and Exascale Computing Jack Dongarra University of Tennessee Oak Ridge National Laboratory University of Manchester 11/17/2010 1 H. Meuer, H. Simon, E. Strohmaier, & JD - Listing of the 500 most powerful Computers in the World - Yardstick: Rmax from LINPACK MPP Ax=b, dense problem TPP performance Rate - Updated twice a year Size SC‘xy in the States in November Meeting in Germany in June - All data available from www.top2 500.org 36rd List: The TOP10 Rmax % of Power Flops/ Rank Site Computer Country Cores [Pflops] Peak [MW] Watt Nat. SuperComputer NUDT YH Cluster, X5670 1 China 186,368 2.57 55 4.04 636 Center in Tianjin 2.93Ghz 6C, NVIDIA GPU DOE / OS Jaguar / Cray 2 USA 224,162 1.76 75 7.0 251 Oak Ridge Nat Lab Cray XT5 sixCore 2.6 GHz Nebulea / Dawning / TC3600 Nat. Supercomputer 3 Blade, Intel X5650, Nvidia China 120,640 1.27 43 2.58 493 Center in Shenzhen C2050 GPU Tusbame 2.0 HP ProLiant GSIC Center, Tokyo 4 SL390s G7 Xeon 6C X5670, Japan 73,278 1.19 52 1.40 850 Institute of Technology Nvidia GPU Hopper, Cray XE6 12-core 5 DOE/SC/LBNL/NERSC USA 153,408 1.054 82 2.91 362 2.1 GHz Commissariat a Tera-100 Bull bullx super- 6 l'Energie Atomique France 138,368 1.050 84 4.59 229 node S6010/S6030 (CEA) DOE / NNSA Roadrunner / IBM 7 USA 122,400 1.04 76 2.35 446 Los Alamos Nat Lab BladeCenter QS22/LS21 NSF / NICS / kyaken/ Cray 8 USA 98,928 .831 81 3.09 269 U of Tennessee Cray XT5 sixCore 2.6 GHz Forschungszentrum Jugene / IBM 9 Germany 294,912 .825 82 2.26 365 Juelich (FZJ) Blue Gene/P Solution DOE/ NNSA / 10 Cray XE6 8-core 2.4 GHz USA 107,152 .817 79 2.95 277 Los Alamos Nat Lab 36rd List: The TOP10 Rmax % of Power Flops/ Rank Site Computer Country Cores [Pflops] Peak [MW] Watt Nat. -
D5.1 Report on Performance and Tuning of Runtime Libraries for ARM
D5.1 Report on performance and tuning of runtime libraries for ARM Architecture Version 1.0 Document Information Contract Number 288777 Project Website www.montblanc-project.eu Contractual Deadline M12 Dissemintation Level PU Nature Report Coordinator Alex Ramirez (BSC) Contributors Isaac Gelado (BSC), Xavier Martorell (BSC), Judit Gimenez (BSC), Harald Servat (BSC), Bernd Mohr (JUELICH), Daniel Lorenz (JUELICH), Marc Schluetter (JUELICH) Reviewers Chris Adeniyi-Jones (ARM) Keywords compiler, runtime system, parallelism, instrumentation, perfor- mance analysis Notices: The research leading to these results has received funding from the European Community's Seventh Framework Programme (FP7/2007-2013) under grant agreement no 288777 c 2011 Mont-Blanc Consortium Partners. All rights reserved. D5.1 - Report on performance and tuning of runtime libraries for ARM Architecture Version 1.0 Change Log Version Description of Change v0.1 Initial draft released to the WP5 contributors v0.2 Draft version sent to internal reviewer v1.0 Final version sent to the EU 2 D5.1 - Report on performance and tuning of runtime libraries for ARM Architecture Version 1.0 Contents Executive Summary 4 1 Introduction 5 2 Runtime Libraries 6 2.1 Mercurium compiler . .6 2.1.1 Porting to ARM-v7 . .6 2.1.2 Fortran support . .6 2.2 Nanos++ . .7 2.3 Extrae instrumentation library . .8 2.4 Scalasca support for the ARM platform . .8 3 Performance Analysis and Tuning 11 4 Conclusions 14 3 D5.1 - Report on performance and tuning of runtime libraries for ARM Architecture Version 1.0 Executive Summary In this Mont-Blanc deliverable we present the current status of porting to the ARM architecture of the OmpSs (Mercurium compiler and Nanos++ runtime system), the Extrae instrumentation library and the Scalasca instrumentation facilities. -
Scalasca User Guide
Scalasca 2.5 User Guide Scalable Automatic Performance Analysis March 2019 The Scalasca Development Team [email protected] The entire code of Scalasca v2 is licensed under the BSD-style license agreement given below, except for the third-party code distributed in the 'vendor/' subdirectory. Please see the corresponding COPYING files in the subdirectories of 'vendor/' included in the distribution tarball for details. Scalasca v2 License Agreement Copyright © 1998–2019 Forschungszentrum Jülich GmbH, Germany Copyright © 2009–2015 German Research School for Simulation Sciences GmbH, Jülich/Aachen, Germany Copyright © 2014–2015 RWTH Aachen University, Germany Copyright © 2003–2008 University of Tennessee, Knoxville, USA Copyright © 2006 Technische Universität Dresden, Germany All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: • Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. • Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. • Neither the names of – the Forschungszentrum Jülich GmbH, – the German Research School for Simulation Sciences GmbH, – the RWTH Aachen University, – the University of Tennessee, – the Technische Universität Dresden, nor the names of their contributors may be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DIS- CLAIMED. -
JSC News No. 166, July 2008
JUMP Succession two more years. This should provide users No. 166 • July 2008 with sufficient available computing power On 1 July 2008, the new IBM Power6 ma- on the general purpose part and time to mi- chine p6 575 – known as JUMP, just like grate their applications to the new cluster. its predecessor – took over the production (Contact: Klaus Wolkersdorfer, ext. 6579) workload from the previous Power4 clus- ter as scheduled. First benchmark tests show that users can expect their applica- Gauss Alliance Founded tions to run about three times faster on the During the ISC 2008, an agreement was new machine. This factor could even be im- signed to found the Gauss Alliance, which proved by making the most out of the follow- will unite supercomputer forces in Ger- ing Power6 features: many. The Gauss Centre for Supercom- • Simultaneous multi-threading (SMT puting (GCS) and eleven regional and topi- mode): 64 threads (instead of 32) can cal high-performance computer centres are be used on one node, where 2 threads participating in the alliance, thus creating a now share one physical processor with computer association that is unique world- its dedicated memory caches and float- wide. The signatories are: Gauss Cen- ing point units. tre for Supercomputing (GCS), Center for • Medium size virtual memory pages Computing and Communication of RWTH (64K): applications can set 64K pages Aachen University, Norddeutscher Ver- as a default during link time and thus bund für Hoch- und Höchstleistungsrech- can benefit from improved hardware ef- nen (HLRN) consisting of Zuse Institute ficiencies by accessing these pages. -
TAU, PAPI, Scalasca and Vampir
Parallel Performance Evaluation Tools: TAU, PAPI, Scalasca and Vampir Two day tutorial at LLNL Building 453 R1001 (Armadillo), May 26-27, Livermore, CA Sameer Shende [email protected] http://www.paratools.com/llnl09 Outline Slide # • Outline and workshop goals 2 • Part I: TAU: A quick reference 8 • Part II: Introduction to performance engineering 77 • Part III: PAPI 105 • Part IV: TAU 119 • Part V: Vampir/VNG 266 • Part VI: Scalasca/KOJAK 307 • Lab Session: PAPI, TAU, Vampir and Scalasca examples 385 2 1 Outline • Day 1: – Introduction to performance evaluation tools: TAU, PAPI, Scalasca, and Vampir – Hands-on: – TAU instrumentation at routine, loop level, PAPI hardware performance counter data collection, derived metrics, analyzing performance using TAU’s paraprof profile browser, using Performance database (PerfDMF), memory evaluation, leak detection • Day 2: – Scalasca, TAU PerfExplorer, VampirServer – Hands-on: – Scalasca bottleneck detection tools, PerfExplorer, trace visualization, workshop examples including the NAS Parallel Benchmarks 3.1 3 Workshop Goals • This tutorial is an introduction to portable performance evaluation tools. • You should leave here with a better understanding of… – Concepts and steps involved in performance evaluation – Understanding key concepts in improving and understanding code performance – How to collect and analyze data from hardware performance counters using PAPI – How to instrument your programs with TAU – Automatic instrumentation at the routine level and outer loop level – Manual instrumentation -
Score-P – a Joint Performance Measurement Run-Time Infrastructure for Periscope, Scalasca, TAU, and Vampir
VIRTUAL INSTITUTE – HIGH PRODUCTIVITY SUPERCOMPUTING Score-P – A Joint Performance Measurement Run-Time Infrastructure for Periscope, Scalasca, TAU, and Vampir VI-HPS Team VIRTUAL INSTITUTE – HIGH PRODUCTIVITY SUPERCOMPUTING Congratulations!? . If you made it this far, you successfully used Score-P to . instrument the application . analyze its execution with a summary measurement, and . examine it with one the interactive analysis report explorer GUIs . ... revealing the call-path profile annotated with . the “Time” metric . Visit counts . MPI message statistics (bytes sent/received) . ... but how good was the measurement? . The measured execution produced the desired valid result . however, the execution took rather longer than expected! . even when ignoring measurement start-up/completion, therefore . it was probably dilated by instrumentation/measurement overhead SCORE-P AND SCALASCA PERFORMANCE TOOLS TRAINING, NERSC, JULY 26, 2016 2 VIRTUAL INSTITUTE – HIGH PRODUCTIVITY SUPERCOMPUTING Performance analysis steps . 0.0 Reference preparation for validation . 1.0 Program instrumentation . 1.1 Summary measurement collection . 1.2 Summary analysis report examination . 2.0 Summary experiment scoring . 2.1 Summary measurement collection with filtering . 2.2 Filtered summary analysis report examination . 3.0 Event trace collection . 3.1 Event trace examination & analysis SCORE-P AND SCALASCA PERFORMANCE TOOLS TRAINING, NERSC, JULY 26, 2016 3 VIRTUAL INSTITUTE – HIGH PRODUCTIVITY SUPERCOMPUTING BT-MZ summary analysis result scoring % scorep-score scorep_bt-mz_B.8x4.<jobid>/profile.cubex . Report scoring as textual Estimated aggregate size of event trace: 40 GB Estimated requirements for largest trace buffer (max_buf): 6 GB output Estimated memory requirements (SCOREP_TOTAL_MEMORY): 6 GB (warning: The memory requirements cannot be satisfied by Score-P to avoid intermediate flushes when tracing. -
Document.Pdf
THE EVOLUTION OF THE HPC FACILITY AT JSC 2019-06-04 I D. KRAUSE (WITH VARIOUS CONTRIBUTIONS) RESEARCH AND DEVELOPMENT @ FZJ on 2.2 Square Kilometres FORSCHUNGSZENTRUM JÜLICH: AT A GLANCE Facts and Figures 1956 Shareholders 11 609.3 5,914 867 90 % Federal Republic FOUNDATION INSTITUTES million euros EMPLOYEES VISITING of Germany REVENUE SCIENTISTS on 12 December 10 % North Rhine- 2 project 2,165 scientists total Westphalia management 536 doctoral from 65 countries organizations researchers (40 % external 323 trainees and funding) students on placement STRATEGIC PRIORITIES CLIMATE RESEARCH QUANTUM COMPUTING LLEC ENERGY STORAGE - SUPER COMPUTING CORE MATERIALS RESEARCH HBP FACILITIES INFORMATION ALZHEIMER’S RESEARCH SOIL BIOECONOMY RESEARCH NEUROMORPHIC COMPUTING - BIO TECHNOLOGY PLANT RESEARCH LARGE-SCALE INSTRUMENTS on campus JÜLICH SUPERCOMPUTING CENTRE JÜLICH SUPERCOMPUTING CENTRE • Supercomputer operation for: • Center – FZJ • Region – RWTH Aachen University • Germany – Gauss Centre for Supercomputing John von Neumann Institute for Computing • Europe – PRACE, EU projects • Application support • Unique support & research environment at JSC • Peer review support and coordination • R-&-D work • Methods and algorithms, computational science, performance analysis and tools • Scientific Big Data Analytics • Computer architectures, Co-Design Exascale Laboratories: EIC, ECL, NVIDIA • Education and Training IBM Power 4+ JUMP, 9 TFlop/s IBM Blue Gene/L IBM Power 6 JUBL, 45 TFlop/s JUMP, 9 TFlop/s JUROPA IBM Blue Gene/P 200 TFlop/s JUGENE, 1 PFlop/s -
Profiling and Tracing Tools for Performance Analysis of Large Scale Applications
Available online at www.prace-ri.eu Partnership for Advanced Computing in Europe Profiling and Tracing Tools for Performance Analysis of Large Scale Applications Jerry Erikssonb, Pedro Ojeda-Mayb, Thomas Ponweisera,*, Thomas Steinreitera aRISC Software GmbH, Softwarepark 35, 4232 Hagenberg, Austria bHigh Performance Computing Center North (HPC2N), MIT Huset, Umeå Universitet, 901 87 Umeå, Sweden Abstract The usage of modern profiling and tracing tools is vital for understanding program behaviour, performance bottlenecks and optimisation potentials in HPC applications. Despite their obvious benefits, such tools are still not that widely adopted within the HPC user community. The two main reasons for this are firstly unawareness and secondly the sometimes inhibitive complexity of getting started with these tools. In this work we aim to address this issue by presenting and comparing the capabilities of four different performance analysis tools, which are 1) HPCToolkit, 2) Extrae and Paraver, 3) SCALASCA and 4) the Intel Trace Analyzer and Collector (ITAC). The practical usage of these tools is demonstrated based on case studies on the widely used molecular dynamics simulation code GROMACS. Keywords: Performance Analysis, Profiling, Tracing, HPCToolkit, Extrae, Paraver, SCALASCA, ITAC, GROMACS 1. Introduction Background of this work is a survey conducted by the PRACE project in the beginning of 2016. In order to develop links to European Centres of Excellence (CoEs) and to understand their needs with respect to high performance computing, a questionnaire has been developed and circulated to each of eight selected CoEs. The results of this survey are summarised in the PRACE-4IP deliverable D7.3 [1]. One finding of this survey which motivated the present work is that from the CoE’s answers it is evident that still the use of profiling tools for code optimization is not that well established. -
Best Practice Guide Modern Processors
Best Practice Guide Modern Processors Ole Widar Saastad, University of Oslo, Norway Kristina Kapanova, NCSA, Bulgaria Stoyan Markov, NCSA, Bulgaria Cristian Morales, BSC, Spain Anastasiia Shamakina, HLRS, Germany Nick Johnson, EPCC, United Kingdom Ezhilmathi Krishnasamy, University of Luxembourg, Luxembourg Sebastien Varrette, University of Luxembourg, Luxembourg Hayk Shoukourian (Editor), LRZ, Germany Updated 5-5-2021 1 Best Practice Guide Modern Processors Table of Contents 1. Introduction .............................................................................................................................. 4 2. ARM Processors ....................................................................................................................... 6 2.1. Architecture ................................................................................................................... 6 2.1.1. Kunpeng 920 ....................................................................................................... 6 2.1.2. ThunderX2 .......................................................................................................... 7 2.1.3. NUMA architecture .............................................................................................. 9 2.2. Programming Environment ............................................................................................... 9 2.2.1. Compilers ........................................................................................................... 9 2.2.2. Vendor performance libraries -
CUBE 4.3.4 – User Guide Generic Display for Application Performance Data
CUBE 4.3.4 – User Guide Generic Display for Application Performance Data April 3, 2016 The Scalasca Development Team [email protected] ii Chapter 0. Copyright Copyright Copyright © 1998–2016 Forschungszentrum Jülich GmbH, Germany Copyright © 2009–2015 German Research School for Simulation Sciences GmbH, Jülich/Aachen, Germany All rights reserved. Redistribution and use in source and binary forms, with or without modification, are per- mitted provided that the following conditions are met: • Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. • Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. • Neither the names of Forschungszentrum Jülich GmbH or German Research School for Simulation Sciences GmbH, Jülich/Aachen, nor the names of their con- tributors may be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIB - UTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITN - ESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDI - NG, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVIC - ES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIB - ILITY OF SUCH DAMAGE. -
An Analysis of System Balance and Architectural Trends Based on Top500 Supercomputers
ORNL/TM-2020/1561 An Analysis of System Balance and Architectural Trends Based on Top500 Supercomputers Hyogi Sim Awais Khan Sudharshan S. Vazhkudai Approved for public release. Distribution is unlimited. August 11, 2020 DOCUMENT AVAILABILITY Reports produced after January 1, 1996, are generally available free via US Department of Energy (DOE) SciTech Connect. Website: www.osti.gov/ Reports produced before January 1, 1996, may be purchased by members of the public from the following source: National Technical Information Service 5285 Port Royal Road Springfield, VA 22161 Telephone: 703-605-6000 (1-800-553-6847) TDD: 703-487-4639 Fax: 703-605-6900 E-mail: [email protected] Website: http://classic.ntis.gov/ Reports are available to DOE employees, DOE contractors, Energy Technology Data Ex- change representatives, and International Nuclear Information System representatives from the following source: Office of Scientific and Technical Information PO Box 62 Oak Ridge, TN 37831 Telephone: 865-576-8401 Fax: 865-576-5728 E-mail: [email protected] Website: http://www.osti.gov/contact.html This report was prepared as an account of work sponsored by an agency of the United States Government. Neither the United States Government nor any agency thereof, nor any of their employees, makes any warranty, express or implied, or assumes any legal lia- bility or responsibility for the accuracy, completeness, or usefulness of any information, apparatus, product, or process disclosed, or rep- resents that its use would not infringe privately owned rights. Refer- ence herein to any specific commercial product, process, or service by trade name, trademark, manufacturer, or otherwise, does not nec- essarily constitute or imply its endorsement, recommendation, or fa- voring by the United States Government or any agency thereof. -
Scalasca 2.3 User Guide Scalable Automatic Performance Analysis
Scalasca 2.3 User Guide Scalable Automatic Performance Analysis May 2016 The Scalasca Development Team [email protected] The entire code of Scalasca v2 is licensed under the BSD-style license agreement given below, except for the third-party code distributed in the 'vendor/' subdirectory. Please see the corresponding COPYING files in the subdirectories of 'vendor/' included in the distribution tarball for details. Scalasca v2 License Agreement Copyright © 1998–2016 Forschungszentrum Jülich GmbH, Germany Copyright © 2009–2015 German Research School for Simulation Sciences GmbH, Jülich/Aachen, Germany Copyright © 2014–2015 RWTH Aachen University, Germany Copyright © 2003–2008 University of Tennessee, Knoxville, USA Copyright © 2006 Technische Universität Dresden, Germany All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: • Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. • Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. • Neither the names of – the Forschungszentrum Jülich GmbH, – the German Research School for Simulation Sciences GmbH, – the RWTH Aachen University, – the University of Tennessee, – the Technische Universität Dresden, nor the names of their contributors may be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DIS- CLAIMED.