SIMD: Data Parallel Execution J

SIMD: Data Parallel Execution J

ERLANGEN REGIONAL COMPUTING CENTER SIMD: Data parallel execution J. Eitzinger PATC, 12.6.2018 Stored Program Computer: Base setting for (int j=0; j<size; j++){ Memory sum = sum + V[j]; } 401d08: f3 0f 58 04 82 addss xmm0,[rdx + rax * 4] Arithmetic Control 401d0d: 48 83 c0 01 add rax,1 Logical 401d11: 39 c7 cmp edi,eax Unit CPU 401d13: 77 f3 ja 401d08 Unit Input Output Architect’s view: Make the common case fast ! Execution and memory . Improvements for relevant software Strategies . What are the technical opportunities? . Increase clock speed . Parallelism . Economical concerns . Specialization . Marketing concerns 2 Data parallel execution units (SIMD) for (int j=0; j<size; j++){ Scalar execution A[j] = B[j] + C[j]; } Register widths • 1 operand = + • 2 operands (SSE) • 4 operands (AVX) • 8 operands (AVX512) 3 Data parallel execution units (SIMD) for (int j=0; j<size; j++){ A[j] = B[j] + C[j]; SIMD execution } Register widths • 1 operand • 2 operands (SSE) = + • 4 operands (AVX) • 8 operands (AVX512) 4 History of short vector SIMD 1995 1996 1997 1998 1999 2001 2008 2010 2011 2012 2013 2016 VIS MDMX MMX 3DNow SSE SSE2 SSE4 VSX AVX IMCI AVX2 AVX512 AltiVec NEON 64 bit 128 bit 256 bit 512 bit Vendors package other ISA features under the SIMD flag: monitor/mwait, NT stores, prefetch ISA, application specific instructions, FMA X86 Power ARM SPARC64 Intel 512bit IBM 128bit A5 upward Fujitsu 256bit AMD 256bit 64bit / 128bit K Computer 128bit 5 Technologies Driving Performance Technology 1991 1992 1993 1994 1995 1996 1997 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008 2009 2010 2011 2012 2013 2014 2015 2016 2017 2018 33 200 1.1 2 3.8 3.2 2.9 2.7 1.9 1.7 Clock MHz MHz GHz GHz GHz GHz GHz GHz GHz GHz ILP SMT SMT2 SMT4 SMT8 SIMD SSE SSE2 AVX AVX512 Multicore 2C 4C 8C 12C 15C 18C 22C 28C 3.2 6.4 12.8 25.6 42.7 60 128 Memory GB/s GB/s GB/s GB/s GB/s GB/s GB/s ILP Obstacle: Not more parallelism available Clock Obstacle: Power/Heat dissipation Multi- Manycore Obstacle: Getting data to/from cores SIMD Obstacle: Power 6 History of Intel chip performance 145W 173W 135W Trade cores for 96W frequency 7 The real picture AVX512 FMA AVX SSE2 8 Finding the right compromise Core complexity Intel Skylake-EP Intel KNL Nvidia GP100 # cores SIMD Area is total power budget! Turbo: Change weights within the same architecture! Frequency 9 AVX and Turbo mode Haswell 2.3 GHz Broadwell 2.3 GHz Turn off Turbo is not an option because base 1456 Xeon E5-2630v4 AVX clock is low! 10 cores 2.2 GHz 10 LINPACK frequency game Highest performance at knee of roofline plot! 11 The driving forces behind performance 2012 Intel IvyBridge-EP P = ncore * F * S * n Intel IvyBridge-EP Number of cores ncore 12 FP instructions per cycle F 2 FP ops per instructions S 4 (DP) / 8 (SP) Clock speed [GHz] n 2.7 Performance [GF/s] P 259 (DP) / 518 (SP) TOP500 rank 1 (1996) But: P=5.4 GF/s for serial, non-SIMD code 12 The driving forces behind performance 2018 Intel Skylake-SP P = ncore * F * M * S * n Intel IvyBridge-EP Number of cores ncore 28 FP instructions per cycle F 2 FMA factor M 2 FP ops per instructions S 8 (DP) / 16 (SP) Clock speed [GHz] n 2.3 (scalar 2.8) Performance [GF/s] P 2060 (DP) / 4122 (SP) But: P=5.6 GF/s for serial, non-SIMD code 13 Future of SIMD in new architectures • For Intel SIMD is a closed chapter • SIMD will be an important aspect of the architecture but not a communicated performance driver • If other vendors go beyond 32b SIMD is questionable • Future developments will include: • Special purpose instructions • Usability improvements • Incremental improvements for existing functionality • What is the next big thing? What about GPGPUs 14 SIMD: THE BASICS SIMD processing – Basics Steps (done by the compiler) for “SIMD processing” for(int i=0; i<n;i++) C[i]=A[i]+B[i]; “Loop unrolling” for(int i=0; i<n;i+=4){ C[i] =A[i] +B[i]; C[i+1]=A[i+1]+B[i+1]; C[i+2]=A[i+2]+B[i+2]; C[i+3]=A[i+3]+B[i+3];} //remainder loop handling Load 256 Bits starting from address of A[i] to LABEL1: register R0 VLOAD R0 A[i] VLOAD R1 B[i] V64ADD[R0,R1] R2 Add the corresponding 64 Bit entries in R0 and VSTORE R2 C[i] R1 and store the 4 results to R2 ii+4 i<(n-4)? JMP LABEL1 Store R2 (256 Bit) to address //remainder loop handling starting at C[i] 16 SIMD processing – Basics No SIMD vectorization for loops with data dependencies: for(int i=0; i<n;i++) A[i]=A[i-1]*s; “Pointer aliasing” may prevent SIMDfication void f(double *A, double *B, double *C, int n) { for(int i=0; i<n; ++i) C[i] = A[i] + B[i]; } C/C++ allows that A &C[-1] and B &C[-2] C[i] = C[i-1] + C[i-2]: dependency No SIMD If “pointer aliasing” is not used, tell it to the compiler: –fno-alias (Intel), -Msafeptr (PGI), -fargument-noalias (gcc) restrict keyword (C only!): void f(double restrict *A, double restrict *B, double restrict *C, int n) {…} 17 Vectorization compiler options (Intel) . The compiler will vectorize starting with –O2. To enable specific SIMD extensions use the –x option: . -xSSE2 vectorize for SSE2 capable machines Available SIMD extensions: SSE2, SSE3, SSSE3, SSE4.1, SSE4.2, AVX, ... -xAVX on Sandy/Ivy Bridge processors . -xCORE-AVX2 on Haswell/Broadwell . -xCORE-AVX512 on Skylake (certain models) . -xMIC-AVX512 on Xeon Phi Knights Landing Recommended option: . -xHost will optimize for the architecture you compile on (Caveat: do not use on standalone KNL, use MIC-AVX512) . To really enable 64b SIMD with current Intel compilers you need to set: -qopt-zmm-usage=high 18 Vectorization source code directives . Fine-grained control of loop vectorization . Use !DEC$ (Fortran) or #pragma (C/C++) sentinel to start a compiler directive . #pragma vector always vectorize even if it seems inefficient (hint!) . #pragma novector do not vectorize even if possible (use with caution) . #pragma vector nontemporal use NT stores if possible (i.e., SIMD and lignment conditions are met) . #pragma vector aligned . specifies that all array accesses are aligned to SIMD address boundaries (DANGEROUS! You must not lie about this!) . Evil interactions with OpenMP possible 20 User mandated vectorization (OpenMP4) . Since Intel Compiler 12.0 the simd pragma is available (now deprecated) . #pragma simd enforces vectorization where the other pragmas fail . Prerequesites: . Countable loop . Innermost loop . Must conform to for-loop style of OpenMP worksharing constructs . There are additional clauses: reduction, vectorlength, private . Refer to the compiler manual for further details #pragma simd reduction(+:x) for (int i=0; i<n; i++) { x = x + A[i]; } . NOTE: Using the #pragma simd the compiler may generate incorrect code if the loop violates the vectorization rules! It may also generate inefficient code! 21 x86 Architecture: SIMD and Alignment . Alignment issues . Alignment of arrays should optimally be on SIMD-width address boundaries to allow packed aligned loads (and NT stores on x86) . Otherwise the compiler will revert to unaligned loads/stores . Modern x86 CPUs have less (not zero) impact for misaligned LOAD/STORE, but Xeon Phi KNC relies heavily on it! . How is manual alignment accomplished? . Stack variables: alignas keyword (C++11/C11) . Dynamic allocation of aligned memory (align = alignment boundary) . C before C11 and C++ before C++17: posix_memalign(void **ptr, size_t align, size_t size); . C11 and C++17: aligned_alloc(size_t align, size_t size); 22 Assembler: Why and how? Why check the assembly code? . Sometimes the only way to make sure the compiler “did the right thing” . Example: “LOOP WAS VECTORIZED” message is printed, but Loads & Stores may still be scalar! . Get the assembler code (Intel compiler): icc –S –O3 -xHost triad.c -o a.out . Disassemble Executable: objdump –d ./a.out | less The x86 ISA is documented in: Intel Software Development Manual (SDM) 2A and 2B AMD64 Architecture Programmer's Manual Vol. 1-5 23 Basics of the x86-64 ISA . Instructions have 0 to 3 operands (4 with AVX-512) . Operands can be registers, memory references or immediates . Opcodes (binary representation of instructions) vary from 1 to 17 bytes . There are two assembler syntax forms: Intel (left) and AT&T (right) . Addressing Mode: BASE + INDEX * SCALE + DISPLACEMENT . C: A[i] equivalent to *(A+i) (a pointer has a type: A+i*8) movaps [rdi + rax*8+48], xmm3 movaps %xmm4, 48(%rdi,%rax,8) add rax, 8 addq $8, %rax js 1b js ..B1.4 401b9f: 0f 29 5c c7 30 movaps %xmm3,0x30(%rdi,%rax,8) 401ba4: 48 83 c0 08 add $0x8,%rax 401ba8: 78 a6 js 401b50 <triad_asm+0x4b> 24 Basics of the x86-64 ISA with extensions 16 general Purpose Registers (64bit): rax, rbx, rcx, rdx, rsi, rdi, rsp, rbp, r8-r15 alias with eight 32 bit register set: eax, ebx, ecx, edx, esi, edi, esp, ebp 8 opmask registers (16bit or 64bit, AVX512 only): k0–k7 Floating Point SIMD Registers: xmm0-xmm15 (xmm31) SSE (128bit) alias with 256-bit and 512-bit registers ymm0-ymm15 (xmm31) AVX (256bit) alias with 512-bit registers zmm0-zmm31 AVX-512 (512bit) SIMD instructions are distinguished by: VEX/EVEX prefix: v Operation: mul, add, mov Modifier: nontemporal (nt), unaligned (u), aligned (a), high (h) Width: scalar (s), packed (p) Data type: single (s), double (d) 25 ISA support on Intel chips SKylake supports all legacy ISA extensions: MMX, SSE, AVX, AVX2 Furthermore KNL supports: .

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    63 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us