Performance Tuning of Scientific Codes with the Roofline Model

Performance Tuning of Scientific Codes with the Roofline Model

Performance Tuning of Scientific Codes with the Roofline Model 1:30pm Introduction to Roofline Samuel Williams 2:00pm Using Roofline in NESAP Jack Deslippe 2:20pm Using LIKWID for Roofline Charlene Yang 2:40pm Using NVProf for Roofline Protonu Basu 3:00pm break / setup NERSC accounts 3:30pm Introduction to Intel Advisor Charlene Yang 3:50pm Hands-on with Intel Advisor Samuel Williams 4:45pm closing remarks / Q&A all Introductions Samuel Williams Jack Deslippe Computational Research Division NERSC Lawrence Berkeley National Lab Lawrence Berkeley National Lab [email protected] [email protected] Charlene Yang Protonu Basu NERSC Computational Research Division Lawrence Berkeley National Lab Lawrence Berkeley National Lab [email protected] [email protected] Introduction to the Roofline Model Samuel Williams Computational Research Division Lawrence Berkeley National Lab [email protected] Acknowledgements § This material is based upon work supported by the Advanced Scientific Computing Research Program in the U.S. Department of Energy, Office of Science, under Award Number DE-AC02-05CH11231. § This material is based upon work supported by the DOE RAPIDS SciDAC Institute. § This research used resources of the National Energy Research Scientific Computing Center (NERSC), which is supported by the Office of Science of the U.S. Department of Energy under contract DE-AC02- 05CH11231. § Special Thanks to: • Zakhar Matveev, Intel Corporation • Roman Belenov, Intel Corporation Introduction to Performance Modeling Why Use Performance Models or Tools? § Identify performance bottlenecks § Motivate software optimizations § Determine when we’re done optimizing • Assess performance relative to machine capabilities • Motivate need for algorithmic changes § Predict performance on future machines / architectures • Sets realistic expectations on performance for future procurements • Used for HW/SW Co-Design to ensure future architectures are well-suited for the computational needs of today’s applications. 6 Performance Models § Many different components can contribute to kernel run time. § Some are characteristics of the application, some are characteristics of the machine, and some are both (memory access pattern + caches). #FP operations Flop/s Cache data movement Cache GB/s DRAM data movement DRAM GB/s PCIe data movement PCIe bandwidth Depth OMP Overhead MPI Message Size Network Bandwidth MPI Send:Wait ratio Network Gap #MPI Wait’s Network Latency 7 Performance Models § Can’t think about all these terms all the time for every application… Computational Complexity #FP operations Flop/s Cache data movement Cache GB/s DRAM data movement DRAM GB/s PCIe data movement PCIe bandwidth Depth OMP Overhead MPI Message Size Network Bandwidth MPI Send:Wait ratio Network Gap #MPI Wait’s Network Latency 8 Performance Models § Because there are so many components, performance models often conceptualize the system as being dominated by one or more of these components. Roofline #FP operations Flop/s Model Cache data movement Cache GB/s DRAM data movement DRAM GB/s PCIe data movement PCIe bandwidth Depth OMP Overhead MPI Message Size Network Bandwidth MPI Send:Wait ratio Network Gap #MPI Wait’s Network Latency Williams et al, "Roofline: An Insightful Visual Performance Model For 9 Multicore Architectures", CACM, 2009. Performance Models § Because there are so many components, performance models often conceptualize the system as being dominated by one or more of these components. #FP operations Flop/s Cache data movement Cache GB/s DRAM data movement DRAM GB/s LogCA PCIe data movement PCIe bandwidth Depth OMP Overhead MPI Message Size Network Bandwidth MPI Send:Wait ratio Network Gap #MPI Wait’s Network Latency Bin Altaf et al, "LogCA: A High-Level Performance Model for Hardware Accelerators", ISCA, 2017. 10 Performance Models § Because there are so many components, performance models often conceptualize the system as being dominated by one or more of these components. #FP operations Flop/s Cache data movement Cache GB/s DRAM data movement DRAM GB/s PCIe data movement PCIe bandwidth Depth OMP Overhead MPI Message Size Network Bandwidth LogGP MPI Send:Wait ratio Network Gap #MPI Wait’s Network Latency Alexandrov, et al, "LogGP: incorporating long messages into the LogP model - one step closer towards a realistic model for parallel 11 computation", SPAA, 1995. Performance Models § Because there are so many components, performance models often conceptualize the system as being dominated by one or more of these components. #FP operations Flop/s Cache data movement Cache GB/s DRAM data movement DRAM GB/s PCIe data movement PCIe bandwidthRight model Depth OMP Overhead MPI Message Size Network Bandwidthdepends! on app and problem size LogP MPI Send:Wait ratio Network Gap #MPI Wait’s Network Latency Culler, et al, "LogP: a practical model of parallel computation", CACM, 1996. 12 Roofline Model: Arithmetic Intensity and Bandwidth Performance Models / Simulators § Historically, many performance models and simulators tracked latencies to predict performance (i.e. counting cycles) § The last two decades saw a number of latency-hiding techniques… • Out-of-order execution (hardware discovers parallelism to hide latency) • HW stream prefetching (hardware speculatively loads data) • Massive thread parallelism (independent threads satisfy the latency-bandwidth product) § Effective latency hiding has resulted in a shift from a latency-limited computing regime to a throughput-limited computing regime 14 Roofline Model § Roofline Model is a throughput- oriented performance model… • Tracks rates not times • Augmented with Little’s Law (concurrency = latency*bandwidth) • Independent of ISA and architecture (applies to CPUs, GPUs, Google TPUs1, etc…) https://crd.lbl.gov/departments/computer-science/PAR/research/roofline 1Jouppi et al, “In-Datacenter Performance Analysis of a Tensor 15 Processing Unit”, ISCA, 2017. (DRAM) Roofline § One could hope to always attain peak performance (Flop/s) CPU § However, finite locality (reuse) and (compute, flop/s) bandwidth limit performance. DRAM Bandwidth § Assume: (GB/s) • Idealized processor/caches DRAM • Cold start (data in DRAM) (data, GB) #FP ops / Peak GFlop/s Time = max #Bytes / Peak GB/s 16 (DRAM) Roofline § One could hope to always attain peak performance (Flop/s) CPU § However, finite locality (reuse) and (compute, flop/s) bandwidth limit performance. DRAM Bandwidth § Assume: (GB/s) • Idealized processor/caches DRAM • Cold start (data in DRAM) (data, GB) Time 1 / Peak GFlop/s = max #FP ops #Bytes / #FP ops / Peak GB/s 17 (DRAM) Roofline § One could hope to always attain peak performance (Flop/s) CPU § However, finite locality (reuse) and (compute, flop/s) bandwidth limit performance. DRAM Bandwidth § Assume: (GB/s) • Idealized processor/caches DRAM • Cold start (data in DRAM) (data, GB) #FP ops Peak GFlop/s = min Time (#FP ops / #Bytes) * Peak GB/s 18 (DRAM) Roofline § One could hope to always attain peak performance (Flop/s) CPU § However, finite locality (reuse) and (compute, flop/s) bandwidth limit performance. DRAM Bandwidth § Assume: (GB/s) • Idealized processor/caches DRAM • Cold start (data in DRAM) (data, GB) Peak GFlop/s GFlop/s = min AI * Peak GB/s Note, Arithmetic Intensity (AI) = Flops / Bytes (as presented to DRAM ) 19 (DRAM) Roofline § Plot Roofline bound using Arithmetic Intensity as the x-axis § Log-log scale makes it easy to Peak Flop/s doodle, extrapolate performance along Moore’s Law, etc… § Kernels with AI less than machine DRAM GB/s balance are ultimately DRAM Attainable Flop/s DRAM-bound Compute-bound bound (we’ll refine this later…) Arithmetic Intensity (Flop:Byte) 20 Roofline Example #1 § Typical machine balance is 5-10 flops per byte… • 40-80 flops per double to exploit compute capability Peak Flop/s • Artifact of technology and money • Unlikely to improve § Consider STREAM Triad… DRAM GB/sGflop/s ≤ AI * DRAM GB/s Attainable Flop/s #pragma omp parallel for for(i=0;i<N;i++){ Z[i] = X[i] + alpha*Y[i]; TRIAD } 0.083 • 2 flops per iteration Arithmetic Intensity (Flop:Byte) • Transfer 24 bytes per iteration (read X[i], Y[i], write Z[i]) • AI = 0.083 flops per byte == Memory bound 21 Roofline Example #2 § Conversely, 7-point constant coefficient stencil… • 7 flops Peak Flop/s • 8 memory references (7 reads, 1 store) per point • Cache can filter all but 1 read and 1 write per point Gflop/s ≤ AI * DRAM GB/s • AI = 0.44 flops per byte == memory bound, but 5x the flop rate #pragma omp parallel for DRAM GB/s for(k=1;k<dim+1;k++){ Attainable Flop/s 7-point for(j=1;j<dim+1;j++){ for(i=1;i<dim+1;i++){ Stencil new[k][j][i] = -6.0*old[k ][j ][i ] TRIAD + old[k ][j ][i-1] + old[k ][j ][i+1] + old[k ][j-1][i ] 0.083 0.44 + old[k ][j+1][i ] Arithmetic Intensity (Flop:Byte) + old[k-1][j ][i ] + old[k+1][j ][i ]; }}} 22 Hierarchical Roofline § Real processors have multiple levels of memory • Registers • L1, L2, L3 cache • MCDRAM/HBM (KNL/GPU device memory) • DDR (main memory) • NVRAM (non-volatile memory) § Applications can have locality in each level § Unique data movements imply unique AI’s § Moreover, each level will have a unique bandwidth 23 Hierarchical Roofline § Construct superposition of Rooflines… § Measure a bandwidth Peak Flop/s § Measure AI for each level of memory • Although an loop nest may have multiple L2 GB/s AI’s and multiple bounds (flops, L1, L2, … DRAM)… MCDRAM cache GB/s DDR Bound DDR AI*BW < • … performance is bound by the Attainable Flop/s DDR GB/s MCDRAM AI*BW minimum Arithmetic Intensity (Flop:Byte) 24 Hierarchical

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    77 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us