SYCL™ 2020 Specification (Revision 3)

Total Page:16

File Type:pdf, Size:1020Kb

SYCL™ 2020 Specification (Revision 3) SYCL™ 2020 Specification (revision 3) The Khronos® SYCL™ Working Group 2021-03-04 21:39:10Z: from git SYCL-2020/final-rev3 commit: 3aa000fbe94ab385edb02d24cadc0b94acd45604 SYCL 2020 rev 3 Table of Contents 1. Acknowledgements . 12 2. Introduction . 14 3. SYCL architecture . 16 3.1. Overview . 16 3.2. Anatomy of a SYCL application . 16 3.3. Normative references . 18 3.4. Non-normative notes and examples . 18 3.5. The SYCL platform model . 18 3.6. The SYCL backend model . 19 3.6.1. Platform mixed version support . 20 3.7. SYCL execution model . 20 3.7.1. SYCL application execution model . 20 3.7.1.1. SYCL backend resources managed by the SYCL application . 21 3.7.1.2. SYCL command groups and execution order . 21 3.7.1.3. Controlling execution order with events . 23 3.7.2. SYCL kernel execution model . 23 3.7.2.1. Basic kernels . 23 3.7.2.2. ND-range kernels . 24 3.7.2.3. Backend-specific kernels . 24 3.8. Memory model . 24 3.8.1. SYCL application memory model . 24 3.8.2. SYCL device memory model . 27 3.8.2.1. Access to memory. 28 3.8.3. SYCL memory consistency model . 28 3.8.3.1. Memory ordering . 29 3.8.3.2. Memory scope. 29 3.8.3.3. Atomic operations . 30 3.8.3.4. Forward progress . 30 3.9. The SYCL programming model . 31 3.9.1. Minimum version of C++ . 31 3.9.2. Alignment with future versions of C++ . 31 3.9.3. Basic data parallel kernels . 31 3.9.4. Work-group data parallel kernels . 32 3.9.5. Hierarchical data parallel kernels. 32 3.9.6. Kernels that are not launched over parallel instances. 33 3.9.7. Pre-defined kernels. 33 3.9.8. Synchronization. 33 3.9.8.1. Synchronization in the SYCL application . 33 3.9.8.2. Synchronization in SYCL kernels . 34 3.9.9. Error handling . 34 3.9.10. Fallback mechanism. 34 2 | Table of Contents SYCL 2020 rev 3 3.9.11. Scheduling of kernels and data movement. 34 3.9.12. Managing object lifetimes . 35 3.9.13. Device discovery and selection . 35 3.9.14. Interfacing with the SYCL backend API . 35 3.10. Memory objects . 36 3.11. Multi-dimensional objects and linearization. 37 3.11.1. Linearization . 37 3.11.2. Multi-dimensional subscript operators . 37 3.12. Implementation options . 37 3.12.1. Single source multiple compiler passes. 38 3.12.2. Single source single compiler pass . 38 3.12.3. Library-only implementation. 38 3.13. Language restrictions in kernels . 38 3.13.1. Device copyable . ..
Recommended publications
  • The Importance of Data
    The landscape of Parallel Programing Models Part 2: The importance of Data Michael Wong and Rod Burns Codeplay Software Ltd. Distiguished Engineer, Vice President of Ecosystem IXPUG 2020 2 © 2020 Codeplay Software Ltd. Distinguished Engineer Michael Wong ● Chair of SYCL Heterogeneous Programming Language ● C++ Directions Group ● ISOCPP.org Director, VP http://isocpp.org/wiki/faq/wg21#michael-wong ● [email protected][email protected] Ported ● Head of Delegation for C++ Standard for Canada Build LLVM- TensorFlow to based ● Chair of Programming Languages for Standards open compilers for Council of Canada standards accelerators Chair of WG21 SG19 Machine Learning using SYCL Chair of WG21 SG14 Games Dev/Low Latency/Financial Trading/Embedded Implement Releasing open- ● Editor: C++ SG5 Transactional Memory Technical source, open- OpenCL and Specification standards based AI SYCL for acceleration tools: ● Editor: C++ SG1 Concurrency Technical Specification SYCL-BLAS, SYCL-ML, accelerator ● MISRA C++ and AUTOSAR VisionCpp processors ● Chair of Standards Council Canada TC22/SC32 Electrical and electronic components (SOTIF) ● Chair of UL4600 Object Tracking ● http://wongmichael.com/about We build GPU compilers for semiconductor companies ● C++11 book in Chinese: Now working to make AI/ML heterogeneous acceleration safe for https://www.amazon.cn/dp/B00ETOV2OQ autonomous vehicle 3 © 2020 Codeplay Software Ltd. Acknowledgement and Disclaimer Numerous people internal and external to the original C++/Khronos group, in industry and academia, have made contributions, influenced ideas, written part of this presentations, and offered feedbacks to form part of this talk. But I claim all credit for errors, and stupid mistakes. These are mine, all mine! You can’t have them.
    [Show full text]
  • Introduction to GPU Computing
    Introduction to GPU Computing J. Austin Harris Scientific Computing Group Oak Ridge National Laboratory ORNL is managed by UT-Battelle for the US Department of Energy Performance Development in Top500 • Yardstick for measuring 1 Exaflop/s performance in HPC – Solve Ax = b – Measure floating-point operations per second (Flop/s) • U.S. targeting Exaflop system as early as 2022 – Building on recent trend of using GPUs https://www.top500.org/statistics/perfdevel 2 J. Austin Harris --- JETSCAPE --- 2020 Hardware Trends • Scaling number of cores/chip instead of clock speed • Power is the root cause – Power density limits clock speed • Goal has shifted to performance through parallelism • Performance is now a software concern Figure from Kathy Yelick, “Ten Ways to Waste a Parallel Computer.” Data from Kunle Olukotun, Lance Hammond, Herb Sutter, Burton Smith, Chris Batten, and Krste Asanoviç. 3 J. Austin Harris --- JETSCAPE --- 2020 GPUs for Computation • Excellent at graphics rendering – Fast computation (e.g., TV refresh rate) – High degree of parallelism (millions of independent pixels) – Needs high memory bandwidth • Often sacrifices latency, but this can be ameliorated • This computation pattern common in many scientific applications 4 J. Austin Harris --- JETSCAPE --- 2020 GPUs for Computation • CPU Strengths • GPU Strengths – Large memory – High mem. BW – Fast clock speeds – Latency tolerant via – Large cache for parallelism latency optimization – More compute – Small number of resources (cores) threads that can run – High perf./watt very quickly • GPU Weaknesses • CPU Weaknesses – Low mem. Capacity – Low mem. bandwidth – Low per-thread perf. – Costly cache misses – Low perf./watt Slide from Jeff Larkin, “Fundamentals of GPU Computing” 5 J.
    [Show full text]
  • Delivering Heterogeneous Programming in C++
    Delivering Heterogeneous Programming in C++ Duncan McBain, Codeplay Software Ltd. About me ● Graduated from Edinburgh University 3 years ago ● Postgrad course got me interested in GPU programming ● Worked at Codeplay since graduating ● Research projects, benchmarking, debuggers ● Most recently on C++ library for heterogeneous systems 2 © 2016 Codeplay Software Ltd. Contents • What are heterogeneous systems? • How can we program them? • The future of heterogeneous systems 3 © 2016 Codeplay Software Ltd. What are heterogeneous systems • By this, I mean devices like GPUs, DSPs, FPGAs… • Generally a bit of hardware that is more specialised than, and fundamentally different to, the host CPU • Specialisation can make it very fast • Can also be harder to program because of specialisation 4 © 2016 Codeplay Software Ltd. Some definitions • Host – The CPU/code that runs on the CPU, controls main memory (RAM), might control many devices • Device – A GPU, DSP, or something more exotic • Heterogeneous system – A host, a device and an API tying them together 5 © 2016 Codeplay Software Ltd. Some definitions • Kernel – Code representing the computation to be performed on the device. • Work group – A collection of many work items executing on a device. Has shared local memory and executes same instructions 6 © 2016 Codeplay Software Ltd. Some definitions ● Work item – A single thread or task on a device that executes in parallel ● Parallel for – Some collection of work items, in many work groups, executing a kernel in parallel. In general, cannot return anything, and must be enqueued asynchronously 7 © 2016 Codeplay Software Ltd. Example heterogeneous device ● CPUs today can execute instructions out-of-order, speculative execution, branch prediction ● Complexity hidden from programmer ● Contrast with e.g.
    [Show full text]
  • Programmers' Tool Chain
    Reduce the complexity of programming multicore ++ Offload™ for PlayStation®3 | Offload™ for Cell Broadband Engine™ | Offload™ for Embedded | Custom C and C++ Compilers | Custom Shader Language Compiler www.codeplay.com It’s a risk to underestimate the complexity of programming multicore applications Software developers are now presented with a rapidly-growing range of different multi-core processors. The common feature of many of these processors is that they are difficult and error-prone to program with existing tools, give very unpredictable performance, and that incompatible, complex programming models are used. Codeplay develop compilers and programming tools with one primary goal - to make it easy for programmers to achieve big performance boosts with multi-core processors, but without needing bigger, specially-trained, expensive development teams to get there. Introducing Codeplay Based in Edinburgh, Scotland, Codeplay Software Limited was founded by veteran games developer Andrew Richards in 2002 with funding from Jez San (the founder of Argonaut Games and ARC International). Codeplay introduced their first product, VectorC, a highly optimizing compiler for x86 PC and PlayStation®2, in 2003. In 2004 Codeplay further developed their business by offering services to processor developers to provide them with compilers and programming tools for their new and unique architectures, using VectorC’s highly retargetable compiler technology. Realising the need for new multicore tools Codeplay started the development of the company’s latest product, the Offload™ C++ Multicore Programming Platform. In October 2009 Offload™: Community Edition was released as a free-to-use tool for PlayStation®3 programmers. Experience and Expertise Codeplay have developed compilers and software optimization technology since 1999.
    [Show full text]
  • Intel® Oneapi Programming Guide
    Intel® oneAPI Programming Guide Intel Corporation www.intel.com Notices and Disclaimers Contents Notices and Disclaimers....................................................................... 5 Chapter 1: Introduction oneAPI Programming Model Overview ..........................................................7 Data Parallel C++ (DPC++)................................................................8 oneAPI Toolkit Distribution..................................................................9 About This Guide.......................................................................................9 Related Documentation ..............................................................................9 Chapter 2: oneAPI Programming Model Sample Program ..................................................................................... 10 Platform Model........................................................................................ 14 Execution Model ...................................................................................... 15 Memory Model ........................................................................................ 17 Memory Objects.............................................................................. 19 Accessors....................................................................................... 19 Synchronization .............................................................................. 20 Unified Shared Memory.................................................................... 20 Kernel Programming
    [Show full text]
  • Opencl SYCL 2.2 Specification
    SYCLTM Provisional Specification SYCL integrates OpenCL devices with modern C++ using a single source design Version 2.2 Revision Date: – 2016/02/15 Khronos OpenCL Working Group — SYCL subgroup Editor: Maria Rovatsou Copyright 2011-2016 The Khronos Group Inc. All Rights Reserved Contents 1 Introduction 13 2 SYCL Architecture 15 2.1 Overview . 15 2.2 The SYCL Platform and Device Model . 15 2.2.1 Platform Mixed Version Support . 16 2.3 SYCL Execution Model . 17 2.3.1 Execution Model: Queues, Command Groups and Contexts . 18 2.3.2 Execution Model: Command Queues . 19 2.3.3 Execution Model: Mapping work-items onto an nd range . 21 2.3.4 Execution Model: Execution of kernel-instances . 22 2.3.5 Execution Model: Hierarchical Parallelism . 24 2.3.6 Execution Model: Device-side enqueue . 25 2.3.7 Execution Model: Synchronization . 25 2.4 Memory Model . 26 2.4.1 Access to memory . 27 2.4.2 Memory consistency . 28 2.4.3 Atomic operations . 28 2.5 The SYCL programming model . 28 2.5.1 Basic data parallel kernels . 30 2.5.2 Work-group data parallel kernels . 30 2.5.3 Hierarchical data parallel kernels . 30 2.5.4 Kernels that are not launched over parallel instances . 31 2.5.5 Synchronization . 31 2.5.6 Error handling . 32 2.5.7 Scheduling of kernels and data movement . 32 2.5.8 Managing object lifetimes . 34 2.5.9 Device discovery and selection . 35 2.5.10 Interfacing with OpenCL . 35 2.6 Anatomy of a SYCL application .
    [Show full text]
  • Of the Threading Building Blocks Flow Graph API, a C++ API for Expressing Dependency, Streaming and Data Flow Applications
    Episode 15: A Proving Ground for Open Standards Host: Nicole Huesman, Intel Guests: Mike Voss, Intel; Andrew Lumsdaine, Northwest Institute for Advanced Computing ___________________________________________________________________________ Nicole Huesman: Welcome to Code Together, an interview series exploring the possibilities of cross- architecture development with those who live it. I’m your host, Nicole Huesman. In earlier episodes, we’ve talked about the need to make it easier for developers to build code for heterogeneous environments in the face of increasingly diverse and data-intensive workloads. And the industry shift to modern C++ with the C++11 release. Today, we’ll continue that conversation, exploring parallelism and heterogeneous computing from a user’s perspective with: Andrew Lumsdaine. As Chief Scientist at the Northwest Institute for Advanced Computing, Andrew wears at least two hats: Laboratory Fellow at Pacific Northwest National Lab, and Affiliate Professor in the Paul G. Allen School of Computer Science and Engineering at the University of Washington. By spanning a university and a national lab, Andrew has the opportunity to work on research questions, and then translate those results into practice. His primary research interest is High Performance Computing, with a particular attention to scalable graph algorithms. I should also note that Andrew is a member of the oneAPI Technical Advisory Board. Andrew, so great to have you with us! Andrew Lumsdaine: Thank you. It’s great to be here. Nicole Huesman: And Mike Voss. Mike is a Principal Engineer at Intel, and was the original architect of the Threading Building Blocks flow graph API, a C++ API for expressing dependency, streaming and data flow applications.
    [Show full text]
  • SYCL – Introduction and Hands-On
    SYCL – Introduction and Hands-on Thomas Applencourt + people on next slide - [email protected] Argonne Leadership Computing Facility Argonne National Laboratory 9700 S. Cass Ave Argonne, IL 60349 Book Keeping Get the example and the presentation: 1 git clone https://github.com/alcf-perfengr/sycltrain 2 cd sycltrain/presentation/2020_07_30_ATPESC People who are here to help: • Collen Bertoni - [email protected] • Brian Homerding - [email protected] • Nevin ”:)” Liber [email protected] • Ben Odom - [email protected] • Dan Petre - [email protected]) 1/17 Table of contents 1. Introduction 2. Theory 3. Hands-on 4. Conclusion 2/17 Introduction What programming model to use to target GPU? • OpenMP (pragma based) • Cuda (proprietary) • Hip (low level) • OpenCL (low level) • Kokkos, RAJA, OCCA (high level, abstraction layer, academic projects) 3/17 What is SYCL™?1 1. Target C++ programmers (template, lambda) 1.1 No language extension 1.2 No pragmas 1.3 No attribute 2. Borrow lot of concept from battle tested OpenCL (platform, device, work-group, range) 3. Single Source (two compilations passes) 4. High level data-transfer 5. SYCL is a Specification developed by the Khronos Group (OpenCL, SPIR, Vulkan, OpenGL) 1SYCL Doesn’t mean ”Someone You Couldn’t Love”. Sadly. 4/17 SYCL Implementations 2 2Credit: Khronos groups (https://www.khronos.org/sycl/) 5/17 Goal of this presentation 1. Give you a feel of SYCL 2. Go through code examples (and make you do some homework) 3. Teach you enough so that can search for the rest if you interested 4. Question are welcomed! 3 3Please just talk, or use slack 6/17 Theory A picture is worth a thousand words4 4and this is a UML diagram so maybe more! 7/17 Memory management: SYCL innovation 1.
    [Show full text]
  • Under the Hood of SYCL – an Initial Performance Analysis with an Unstructured-Mesh CFD Application
    Under the Hood of SYCL – An Initial Performance Analysis With an Unstructured-mesh CFD Application Istvan Z. Reguly1;2, Andrew M.B. Owenson2, Archie Powell2, Stephen A. Jarvis3, and Gihan R. Mudalige2 1 Faculty of Information Technology and Bionics, Pazmany Peter Catholic University, Budapest, Hungary [email protected] 2 University of Warwick, Coventry, UK {a.m.b.owenson, a.powell.3, g.mudalige}@warwick.ac.uk 3 University of Birmingham, Birmingham, UK [email protected] Abstract. As the computing hardware landscape gets more diverse, and the com- plexity of hardware grows, the need for a general purpose parallel programming model capable of developing (performance) portable codes have become highly attractive. Intel’s OneAPI suite, which is based on the SYCL standard aims to fill this gap using a modern C++ API. In this paper, we use SYCL to parallelize MG- CFD, an unstructured-mesh computational fluid dynamics (CFD) code, to explore current performance of SYCL. The code is benchmarked on several modern pro- cessor systems from Intel (including CPUs and the latest Xe LP GPU), AMD, ARM and Nvidia, making use of a variety of current SYCL compilers, with a par- ticular focus on OneAPI and how it maps to Intel’s CPU and GPU architectures. We compare performance with other parallelisations available in OP2, including SIMD, OpenMP, MPI and CUDA. The results are mixed; the performance of this class of applications, when parallelized with SYCL, highly depends on the target architecture and the compiler, but in many cases comes close to the performance of currently prevalent parallel programming models.
    [Show full text]
  • Resyclator: Transforming CUDA C++ Source Code Into SYCL
    ReSYCLator: Transforming CUDA C++ source code into SYCL Tobias Stauber Peter Sommerlad [email protected] [email protected] IFS Institute for Software at FHO-HSR Hochschule für Technik Rapperswil, Switzerland Figure 1: Transforming CUDA C++ code to SYCL using C++ AST Rewriting. ABSTRACT ReSYCLator is a plug-in for the C++ IDE Cevelop[2], that is CUDA™ while very popular, is not as flexible with respect to tar- itself an extension of Eclipse-CDT. ReSYCLator bridges the gap get devices as OpenCL™. While parallel algorithm research might between algorithm availability and portability, by providing au- address problems first with a CUDA C++ solution, those results are tomatic transformation of CUDA C++ code to SYCL C++. A first ® not easily portable to a target not directly supported by CUDA. In attempt basing the transformation on NVIDIA ’s Nsight™ Eclipse contrast, a SYCL™ C++ solution can operate on the larger variety CDT plug-in showed that Nsight™’s weak integration into CDT’s of platforms supported by OpenCL. static analysis and refactoring infrastructure is insufficient. There- fore, an own CUDA-C++ parser and CUDA language support for Eclipse CDT was developed (CRITTER) that is a sound platform Permission to make digital or hard copies of all or part of this work for personal or for transformations from CUDA C++ programs to SYCL based on classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation AST transformations. on the first page. Copyrights for components of this work owned by others than the author(s) must be honored.
    [Show full text]
  • An Introduction to GPU Computing
    An Introduction to GPU Computing iVEC Supercomputer Training 5th - 9th November 2012 Introducing the Historical GPU Graphics Processing Unit (GPU) n : A specialised electronic circuit designed to rapidly manipulate and alter memory in such a way as to accelerate the building of images in a frame buffer intended for output to a display. Introducing the Modern GPU Graphics Processing Unit (GPU) n : A general purpose electronic circuit designed to rapidly manipulate and alter memory in such a way as to accelerate computational algorithms that have fine-grained parallelism. GPU Computing Motivation But what about... Central Processing Unit (CPU) n : the portion of a computer system that carries out the instructions of a computer program, to perform the basic arithmetical, logical, and input/output operations of the system GPU Computing Motivation : Arithmetic Performance GPU Computing Motivation How is that possible? - GPUs have less cache and logic, more arithmetic - GPUs execute lots of simple threads - GPUs are physically bigger Nehalem Die (263mm2) Fermi Die (530mm2) GPU Computing Motivation : Power Efficiency GPU Computing Motivation : Memory Bandwidth GPU Computing Motivation : Summary Pros : - high arithmetic performance - high memory bandwidth - power efficient computation Cons : - separated by PCI Express bus - requires fine grain parallelism - requires high arithmetic intensity - requires light weight threads - requires additional software development For the right algorithm, additional time spent on code development can yield significant
    [Show full text]
  • SYCL Building Blocks for C++ Libraries Maria Rovatsou, SYCL Specification Editor Principal Software Engineer, Codeplay Software
    SYCL Building Blocks for C++ libraries Maria Rovatsou, SYCL specification editor Principal Software Engineer, Codeplay Software © Copyright Khronos Group 2014 SYCL Building Blocks for C++ libraries Overview SYCL standard What is the SYCL toolkit for? C++ libraries using SYCL: Under construction SYCL Building blocks asynchronous execution on multiple SYCL devices command groups for embedding computational kernels in a library buffers and accessors for data storage and access vector classes shared between host and device pipes hierarchical command group execution printing on device using cl::sycl::stream © Copyright Khronos Group 2014 SYCL standard © Copyright Khronos Group 2014 Khronos royalty-free open standard that works with OpenCLTM SYCLTM for OpenCLTM 1.2 is a published specification that can be found at: https://www.khronos.org/registry/sycl/specs/sycl-1.2.pdf SYCLTM for OpenCLTM 2.2 is a published provisional specification that the group would like feedback on: https://www.khronos.org/registry/sycl/specs/sycl-2.2.pdf © Copyright Khronos Group 2014 What is the SYCL toolkit for? © Copyright Khronos Group 2014 What is the SYCL toolkit for? Performance Heterogeneous Programmability Power efficiency system Longevity of software written for Best h/w architecture integrating host multi-core CPU custom architectures per application with GPUs and Use existing C++ frameworks on Harness processing power in accelerators new targets, e.g. from HPC to order to create much more mobile and embedded and vice powerful applications versa © Copyright
    [Show full text]