Clmagma: High Performance Dense Linear Algebra with Opencl∗

Total Page:16

File Type:pdf, Size:1020Kb

Clmagma: High Performance Dense Linear Algebra with Opencl∗ clMAGMA: High Performance Dense Linear Algebra with OpenCL∗ y z Chongxiao Cao Jack Dongarra Peng Du Innovative Computing Innovative Computing Amazon.com Laboratory Laboratory Seattle, WA University of Tennessee University of Tennessee [email protected] Knoxville, TN Knoxville, TN [email protected] [email protected] Mark Gates Piotr Luszczek Stanimire Tomov Innovative Computing Innovative Computing Innovative Computing Laboratory Laboratory Laboratory University of Tennessee University of Tennessee University of Tennessee Knoxville, TN Knoxville, TN Knoxville, TN [email protected] [email protected] [email protected] ABSTRACT Categories and Subject Descriptors This paper presents the design and implementation of sev- G.4 [Mathematical software]: Algorithm design and anal- eral fundamental dense linear algebra (DLA) algorithms in ysis, Efficiency, Parallel implementations, Portability; G.1.3 OpenCL. In particular, these are linear system solvers and [Numerical analysis]: Numerical linear algebra|linear eigenvalue problem solvers. Further, we give an overview of systems, matrix inversion, eigenvalues and eigenvectors the clMAGMA library, an open source, high performance OpenCL library that incorporates the developments pre- 1. INTRODUCTION sented, and in general provides to heterogeneous architec- Solving linear systems of equations and eigenvalue problems tures the DLA functionality of the popular LAPACK library. is fundamental to scientific computing. The popular LA- The LAPACK-compliance and use of OpenCL simplify the PACK library [5], and in particular its vendor optimized use of clMAGMA in applications, while providing them implementations like Intel's MKL [13] or AMD's ACML [3], with portably performant DLA. High performance is ob- have been the libraries of choice to provide these solvers tained through use of the high-performance OpenCL BLAS, for dense matrices on shared memory systems. This pa- hardware and OpenCL-specific tuning, and a hybridization per considers a redesign of the LAPACK algorithms and methodology where we split the algorithm into computa- their OpenCL implementation to add efficient support for tional tasks of various granularities. Execution of those heterogeneous systems of multicore processors with GPU tasks is properly scheduled over the heterogeneous hardware accelerators and coprocessors. This is not the first time components by minimizing data movements and mapping that DLA libraries have needed a redesign to be efficient algorithmic requirements to the architectural strengths of on new architectures { notable examples being the move the various heterogeneous hardware components. from LINPACK [10] to LAPACK [5] in the 80's to make algorithms cache friendly, ScaLAPACK [8] in the 90's to ∗This research was sponsored by the National Science Foun- support distributed memory systems, and now the PLASMA dation through the Keeneland: National Institute for Experi- and MAGMA libraries [1] targeting efficiency on multicore mental Computing grant (award #0910735), the Department and heterogeneous architectures, respectively. of Energy, and AMD. y Also affiliated with the Oak Ridge National Laboratory, TN, The development of new high-performance numerical libraries USA and the University of Manchester, UK is complex, accounting for the extreme level of parallelism, zResearch completed while at the University of Tennessee, heterogeneity, and wide variety of accelerators and coproces- Knoxville sors available in current architectures. Challenges vary from new algorithmic designs to choices of programming models, languages, and frameworks that ease development, future maintenance, and portability. This paper addresses these issues while presenting our approach and algorithmic designs in the development of the clMAGMA [9] library. To provide a uniform portability across a variety of GPU ac- celerators and coprocessors (e.g., Intel Xeon Phi), clMAGMA uses OpenCL [14]. OpenCL is an open standard for off- loading computations to accelerators, coprocessors, and multi/ manycore processors, and is maintained by the Khronos group Currently, the most complete OpenCL BLAS implementa- with the backing of major hardware and software computer tion is AMD's clAmdBlas, provided through the AMD's industry vendors. It offers portability across hardware and Accelerated Parallel Processing Math Libraries (APPML) OS software. Although the use of OpenCL provides program- [2]. It can be used on architectures other than AMD, but its ming portability, cross-device performance portability is not tuning, and therefore highest efficiency, is on AMD hardware. guaranteed; we specifically address this in Section 2. The potential of OpenCL to express BLAS algorithms (vs. other, lower level access to the hardware languages) while To deal with the extreme level of parallelism and heterogene- obtaining high performance is evident through the clAmd- ity in current architectures, clMAGMA uses a hybridization Blas. Other implementations, e.g., from Nakasato et al. [16, methodology, described in Section 3, where we split the 15], confirm this by obtaining impressive high performance algorithms of interest into computational tasks of various matrix-matrix multiplication (GEMM). In particular, the granularities, and properly schedule those tasks' execution highest performance that we are aware of has been demon- over the heterogeneous hardware. Thus, we use a Directed strated by Matsumoto et al. [15] { their OpenCL DGEMM Acyclic Graph (DAG) approach to parallelism and schedul- reaches up to 848 Gflop/s, and SGEMM up to 2,646 Gflop/s, ing that has been developed and successfully used for dense which is 90% and 70% of the double and single precision linear algebra libraries such as PLASMA and MAGMA [1], peak, respectively, of AMD's Tahiti GPU (Radeon HD 7970). as well as in general task-based approaches to parallelism, such as runtime systems like StarPU [6] and SMPSs [7]. In previous work, we evaluated OpenCL as a programming tool for performance-portable BLAS [11]. Triangular solvers Besides the general cross-device considerations addressed in (TRSM) and GEMMs were developed in OpenCL, tuned for Section 2, obtaining high performance in OpenCL depends a specific device, and compared. The conclusion was that on a combination of algorithm and hardware-specific opti- OpenCL environment setup overhead is large and should mizations, discussed in Section 4. The implication of this be minimized, e.g., by preprocessing or localized in library on software, in order to maintain its performance portabil- initialization routines. More importantly, the performance re- ity across hardware, is the need to build in it algorithmic sults presented confirmed the conclusion above that OpenCL variations that are tunable, e.g., at installation time. This is expressive enough for developing high performance BLAS, is the basis of autotuning, an example of these advanced so long as architectural specifics are taken into account in the optimization techniques. algorithm design. Even though good performance should not be expected from blindly running algorithms on a new plat- A performance study on AMD hardware is presented in Sec- form, autotuning heuristics can help to improve performance tion 5. Besides verifying our approaches and confirming the on a single platform. appeal of OpenCL and accelerators for high-performance DLA, the results open up a number of future work opportu- Autotuning mechanisms are already provided in clAmdBlas nities discussed in our conclusions. through a tuning tool that the user can run to produce optimized OpenCL BLAS on the architecture of interest. 2. CROSS-DEVICE CONSIDERATIONS Thus, as performance portability of OpenCL BLAS can be A recommended approach to developing a high-performance obtained, organizing higher-level libraries like clMAGMA and easy to maintain DLA library is to express the algorithms in terms of OpenCL BLAS can ensure their performance of interest in terms of the BLAS standard. Performance portability as well. portability is then obtained through the use of architecture- specific, highly tuned BLAS implementations (e.g., MKL 2.2 Microbenchmarks from Intel or ACML from AMD). LAPACK and ScaLAPACK We developed a number of microbenchmarks to help us gain a have demonstrated this over the years, and now we see it in better understanding of OpenCL and to guide our algorithm the new MAGMA and PLASMA libraries. The clMAGMA design and tuning. We describe two benchmarks that can be library takes the same approach, and therefore performance key for performance { kernel launch overhead and CPU-GPU portability relies on the availability of portable OpenCL data transfer. To add some context to the measurements BLAS, discussed in Section 2.1. Specifics related to OpenCL reported, we include comparisons with corresponding CUDA and its implementation are also important for obtaining high- measurements. performance and must be addressed while designing and tuning OpenCL algorithms. Well designed microbenchmarks, 2.2.1 Kernel launch overhead shown in Section 2.2, can be used to obtain these key OpenCL The average time to asynchronously invoke an OpenCL 1.2 specifics to achieving high performance. AMD-APP (1016.4) kernel on an AMD Tahiti GPU (Radeon HD 7900 Series) is 1:0{1:5µs. This was measured by asyn- 2.1 Portable OpenCL BLAS chronously launching an empty kernel a large number of The Automatically Tuned Linear Algebra Software (ATLAS) times and synchronizing at the end. The overhead increases library [19] is a BLAS implementation for CPUs. ATLAS to 120µs when synchronizing
Recommended publications
  • (LINUX) Computer in Three Parts
    Science on a (LINUX) computer in three parts Introductory short couse, Part II Thorsten Becker University of Southern California, Los Angeles September 2007 The first part dealt with ● UNIX: what and why ● File system, Window managers ● Shell environment ● Editing files ● Command line tools ● Scripts and GUIs ● Virtualization Contents part II ● Typesetting ● Programming – common languages – philosophy – compiling, debugging, make, version control – C and F77 interfacing – libraries and packages ● Number crunching ● Visualization tools Programming: Traditional languages in the natural sciences ● Fortran: higher level, good for math – F77: legacy, don't use (but know how to read) – F90/F95: nice vector features, finally implements C capabilities (structures, memory allocation) ● C: low level (e.g. pointers), better structured – very close to UNIX philosophy – structures offer nice way of modular programming, see Wikipedia on C ● I recommend F95, and use C happily myself Programming: Some Languages that haven't completely made it to scientific computing ● C++: object oriented programming model – reusable objects with methods and such – can be partly realized by modular programming in C ● Java: what's good for commercial projects (or smart, or elegant) doesn't have to be good for scientific computing ● Concern about portability as well as general access Programming: Compromises ● Python – Object oriented – Interpreted – Interfaces easily with F90/C – Numerous scientific packages Programming: Other interpreted, high- abstraction languages
    [Show full text]
  • Some Notes on BLAS, SIMD, MIC and GPGPU (For Octave and Matlab)
    Some Notes on BLAS, SIMD, MIC and GPGPU (for Octave and Matlab) Christian Himpe ([email protected]) WWU Münster Institute for Computational and Applied Mathematics Software-Tools für die Numerische Mathematik 15.10.2014 About1 BLAS & LAPACK Affinity SIMD Automatic Offloading 1 Get your Buzzword-Bingo ready! BLAS & LAPACK BLAS (Basic Linear Algebra System) Level 1: vector-vector operations (dot-product, vector norms, generalized vector addition) Level 2: matrix-vector operations (generalized matrix-vector multiplication) Level 3: matrix-matrix operations (generalized matrix-matrix multiplication) LAPACK (Linear Algebra Package) Matrix Factorizations: LU, QR, Cholesky, Schur Least-Squares: LLS, LSE, GLM Eigenproblems: SEP, NEP, SVD Default (un-optimized) netlib reference implementation. Used by Octave, Matlab, SciPy/NumPy (Python), Julia, R. MKL Intel’s implementation of BLAS and LAPACK. MKL (Math Kernel Library) can offload computations to XeonPhi can use OpenMP provides additionally FFT, libm and 1D-interpolation Current Version: 11.0.5 (automatic offloading to Phis) Costs (we have it, and Matlab ships with it, too) Go to: https://software.intel.com/en-us/intel-mkl ACML AMD doesn’t want to be left out. ACML (AMD Core Math Libraries) Can use OpenCL for automatic offloading to GPUs Special version to exploit FMA(4) instructions Choice of compiled binaries (GFortran, Intel Fortran, Open64, PGI) Current Version: 6 (automatic offloading to GPUs) Free (but not open source) Go to: http://developer.amd.com/tools-and-sdks/ cpu-development/amd-core-math-library-acml OpenBLAS Open Source, the third kind. OpenBLAS Fork of GotoBLAS Good performance for dense operations (close to the MKL) Can use OpenMP and is compiled for specific architecture Current Version: 0.2.11 Open source (!) Go to: http://github.com/xianyi/OpenBLAS FlexiBLAS So many BLAS implementations, so little time..
    [Show full text]
  • Parallel Programming Models for Dense Linear Algebra on Heterogeneous Systems
    DOI: 10.14529/jsfi150405 Parallel Programming Models for Dense Linear Algebra on Heterogeneous Systems M. Abalenkovs1, A. Abdelfattah2, J. Dongarra 1,2,3, M. Gates2, A. Haidar2, J. Kurzak2, P. Luszczek2, S. Tomov2, I. Yamazaki2, A. YarKhan2 c The Authors 2015. This paper is published with open access at SuperFri.org We present a review of the current best practices in parallel programming models for dense linear algebra (DLA) on heterogeneous architectures. We consider multicore CPUs, stand alone manycore coprocessors, GPUs, and combinations of these. Of interest is the evolution of the programming models for DLA libraries – in particular, the evolution from the popular LAPACK and ScaLAPACK libraries to their modernized counterparts PLASMA (for multicore CPUs) and MAGMA (for heterogeneous architectures), as well as other programming models and libraries. Besides providing insights into the programming techniques of the libraries considered, we outline our view of the current strengths and weaknesses of their programming models – especially in regards to hardware trends and ease of programming high-performance numerical software that current applications need – in order to motivate work and future directions for the next generation of parallel programming models for high-performance linear algebra libraries on heterogeneous systems. Keywords: Programming models, runtime, HPC, GPU, multicore, dense linear algebra. Introduction Parallelism in today’s computer architectures is pervasive – not only in systems from large supercomputers to laptops, but also in small portable devices like smart phones and watches. Along with parallelism, the level of heterogeneity in modern computing systems is also gradually increasing. Multicore CPUs are often combined with discrete high-performance accelerators like graphics processing units (GPUs) and coprocessors like Intel’s Xeon Phi, but are also included as integrated parts with system-on-chip (SoC) platforms, as in the AMD Fusion family of ap- plication processing units (APUs) or the NVIDIA Tegra mobile family of devices.
    [Show full text]
  • AMD Core Math Library (ACML)
    AMD Core Math Library (ACML) Version 4.2.0 Copyright c 2003-2008 Advanced Micro Devices, Inc., Numerical Algorithms Group Ltd. AMD, the AMD Arrow logo, AMD Opteron, AMD Athlon and combinations thereof are trademarks of Advanced Micro Devices, Inc. i Short Contents 1 Introduction................................... 1 2 General Information ............................. 2 3 BLAS: Basic Linear Algebra Subprograms ............. 19 4 LAPACK: Package of Linear Algebra Subroutines ........ 20 5 Fast Fourier Transforms (FFTs) .................... 24 6 Random Number Generators....................... 75 7 ACML MV: Fast Math and Fast Vector Math Library .... 163 8 References .................................. 229 Subject Index ................................... 230 Routine Index ................................... 233 ii Table of Contents 1 Introduction ............................... 1 2 General Information ....................... 2 2.1 Determining the best ACML version for your system ........... 2 2.2 Accessing the Library (Linux) ................................ 4 2.2.1 Accessing the Library under Linux using GNU gfortran/gcc ......................................................... 4 2.2.2 Accessing the Library under Linux using PGI compilers pgf77/pgf90/pgcc ......................................... 5 2.2.3 Accessing the Library under Linux using PathScale compilers pathf90/pathcc ........................................... 6 2.2.4 Accessing the Library under Linux using the NAGWare f95 compiler.................................................
    [Show full text]
  • 35.232-2016.43 Lamees Elhiny.Pdf
    LOAD PARTITIONING FOR MATRIX-MATRIX MULTIPLICATION ON A CLUSTER OF CPU-GPU NODES USING THE DIVISIBLE LOAD PARADIGM by Lamees Elhiny A Thesis Presented to the Faculty of the American University of Sharjah College of Engineering in Partial Fulfillment of the Requirements for the Degree of Master of Science in Computer Engineering Sharjah, United Arab Emirates November 2016 c 2016. Lamees Elhiny. All rights reserved. Approval Signatures We, the undersigned, approve the Master’s Thesis of Lamees Elhiny Thesis Title: Load Partitioning for Matrix-Matrix Multiplication on a Cluster of CPU- GPU Nodes Using the Divisible Load Paradigm Signature Date of Signature (dd/mm/yyyy) ___________________________ _______________ Dr. Gerassimos Barlas Professor, Department of Computer Science and Engineering Thesis Advisor ___________________________ _______________ Dr. Khaled El-Fakih Associate Professor, Department of Computer Science and Engineering Thesis Committee Member ___________________________ _______________ Dr. Naoufel Werghi Associate Professor, Electrical and Computer Engineering Department Khalifa University of Science, Technology & Research (KUSTAR) Thesis Committee Member ___________________________ _______________ Dr. Fadi Aloul Head, Department of Computer Science and Engineering ___________________________ _______________ Dr. Mohamed El-Tarhuni Associate Dean, College of Engineering ___________________________ _______________ Dr. Richard Schoephoerster Dean, College of Engineering ___________________________ _______________ Dr. Khaled Assaleh Interim Vice Provost for Research and Graduate Studies Acknowledgements I would like to express my sincere gratitude to my thesis advisor, Dr. Gerassi- mos Barlas, for his guidance at every step throughout the thesis. I am thankful to Dr. Assim Sagahyroon, and to the Department of Computer Science & Engineering as well as the American University of Sharjah for offering me the Graduate Teaching Assistantship, which allowed me to pursue my graduate studies.
    [Show full text]
  • Parallel Siesta.Graffle
    Siesta in parallel Toby White Dept. Earth Sciences, University of Cambridge ❖ How to build Siesta in parallel ❖ How to run Siesta in parallel ❖ How Siesta works in parallel ❖ How to use parallelism efficiently ❖ How to build Siesta in parallel Siesta SCALAPACK BLACS LAPACK MPI BLAS LAPACK Vector/matrix BLAS manipulation Linear Algebra PACKage http://netlib.org/lapack/ Basic Linear Algebra Subroutines http://netlib.org/blas/ ATLAS BLAS: http://atlas.sf.net Free, open source (needs separate LAPACK) GOTO BLAS: http://www.tacc.utexas.edu/resources/software/#blas Free, registration required, source available (needs separate LAPACK) Intel MKL (Math Kernel Library): Intel compiler only Not cheap (but often installed on supercomputers) ACML (AMD Core Math Library) http://developer.amd.com/acml.jsp Free, registration required. Versions for most compilers. Sun Performance Library http://developers.sun.com/sunstudio/perflib_index.html Free, registration required. Only for Sun compilers (Linux/Solaris) IBM ESSL (Engineering & Science Subroutine Library) http://www-03.ibm.com/systems/p/software/essl.html Free, registration required. Only for IBM compilers (Linux/AIX) Parallel MPI communication Message Passing Infrastructure http://www.mcs.anl.gov/mpi/ You probably don't care - your supercomputer will have (at least one) MPI version installed. Just in case, if you are building your own cluster: MPICH2: http://www-unix.mcs.anl.gov/mpi/mpich2/ Open-MPI: http://www.open-mpi.org And a lot of experimental super-fast versions ... SCALAPACK BLACS Parallel linear algebra Basic Linear Algebra Communication Subprograms http://netlib.org/blacs/ SCAlable LAPACK http://netlib.org/scalapack/ Intel MKL (Math Kernel Library): Intel compiler only AMCL (AMD Core Math Library) http://developer.amd.com/acml.jsp S3L (Sun Scalable Scientific Program Library) http://www.sun.com/software/products/clustertools/ IBM PESSL (Parallel Engineering & Science Subroutine Library) http://www-03.ibm.com/systems/p/software/essl.html Previously mentioned libraries are not the only ones - just most common.
    [Show full text]
  • AMD Core Math Library (ACML) ©
    ¨ AMD Core Math Library (ACML) © Version 5.3.1 Copyright c 2003-2013 Advanced Micro Devices, Inc., Numerical Algorithms Group Ltd. All rights reserved. The contents of this document are provided in connection with Advanced Micro Devices, Inc. (\AMD") products. AMD makes no representations or warranties with respect to the accuracy or completeness of the contents of this publication and reserves the right to make changes to specifications and product descriptions at any time without notice. The information contained herein may be of a preliminary or advance nature and is subject to change without notice. No license, whether express, implied, arising by estoppel, or oth- erwise, to any intellectual property rights are granted by this publication. Except as set forth in AMD's Standard Terms and Conditions of Sale, AMD assumes no liability whatso- ever, and disclaims any express or implied warranty, relating to its products including, but not limited to, the implied warranty of merchantability, fitness for a particular purpose, or infringement of any intellectual property right. AMD's products are not designed, intended, authorized or warranted for use as components in systems intended for surgical implant into the body, or in other applications intended to support or sustain life, or in any other application in which the failure of AMD's product could create a situation where personal injury, death, or severe property or environmental damage may occur. AMD reserves the right to discontinue or make changes to its products at any time without notice. Trademarks AMD, the AMD Arrow logo, and combinations thereof, AMD Athlon, and AMD Opteron are trademarks of Advanced Micro Devices, Inc.
    [Show full text]
  • AMD Optimizing CPU Libraries User Guide Version 2.1
    [AMD Public Use] AMD Optimizing CPU Libraries User Guide Version 2.1 [AMD Public Use] AOCL User Guide 2.1 Table of Contents 1. Introduction .......................................................................................................................................... 4 2. Supported Operating Systems and Compliers ...................................................................................... 5 3. BLIS library for AMD .............................................................................................................................. 6 3.1. Installation ........................................................................................................................................ 6 3.1.1. Build BLIS from source .................................................................................................................. 6 3.1.1.1. Single-thread BLIS ..................................................................................................................... 6 3.1.1.2. Multi-threaded BLIS .................................................................................................................. 6 3.1.2. Using pre-built binaries ................................................................................................................. 7 3.2. Usage ................................................................................................................................................. 7 3.2.1. BLIS Usage in FORTRAN ................................................................................................................
    [Show full text]
  • Best Practice Guide - Generic X86 Vegard Eide, NTNU Nikos Anastopoulos, GRNET Henrik Nagel, NTNU 02-05-2013
    Best Practice Guide - Generic x86 Vegard Eide, NTNU Nikos Anastopoulos, GRNET Henrik Nagel, NTNU 02-05-2013 1 Best Practice Guide - Generic x86 Table of Contents 1. Introduction .............................................................................................................................. 3 2. x86 - Basic Properties ................................................................................................................ 3 2.1. Basic Properties .............................................................................................................. 3 2.2. Simultaneous Multithreading ............................................................................................. 4 3. Programming Environment ......................................................................................................... 5 3.1. Modules ........................................................................................................................ 5 3.2. Compiling ..................................................................................................................... 6 3.2.1. Compilers ........................................................................................................... 6 3.2.2. General Compiler Flags ......................................................................................... 6 3.2.2.1. GCC ........................................................................................................ 6 3.2.2.2. Intel ........................................................................................................
    [Show full text]
  • Accelerating Dense Linear Algebra for Gpus, Multicores and Hybrid Architectures: an Autotuned and Algorithmic Approach
    University of Tennessee, Knoxville TRACE: Tennessee Research and Creative Exchange Masters Theses Graduate School 8-2010 Accelerating Dense Linear Algebra for GPUs, Multicores and Hybrid Architectures: an Autotuned and Algorithmic Approach Rajib Kumar Nath [email protected] Follow this and additional works at: https://trace.tennessee.edu/utk_gradthes Recommended Citation Nath, Rajib Kumar, "Accelerating Dense Linear Algebra for GPUs, Multicores and Hybrid Architectures: an Autotuned and Algorithmic Approach. " Master's Thesis, University of Tennessee, 2010. https://trace.tennessee.edu/utk_gradthes/734 This Thesis is brought to you for free and open access by the Graduate School at TRACE: Tennessee Research and Creative Exchange. It has been accepted for inclusion in Masters Theses by an authorized administrator of TRACE: Tennessee Research and Creative Exchange. For more information, please contact [email protected]. To the Graduate Council: I am submitting herewith a thesis written by Rajib Kumar Nath entitled "Accelerating Dense Linear Algebra for GPUs, Multicores and Hybrid Architectures: an Autotuned and Algorithmic Approach." I have examined the final electronic copy of this thesis for form and content and recommend that it be accepted in partial fulfillment of the equirr ements for the degree of Master of Science, with a major in Computer Science. Jack Dongarra, Major Professor We have read this thesis and recommend its acceptance: Stanimire Z. Tomov, Lynne E. Parker Accepted for the Council: Carolyn R. Hodges Vice Provost and Dean of the Graduate School (Original signatures are on file with official studentecor r ds.) To the Graduate Council: I am submitting herewith a thesis written by Rajib Kumar Nath entitled \Accel- erating Dense Linear Algebra for GPUs, Multicores and Hybrid Architectures: an Autotuned and Algorithmic Approach." I have examined the final paper copy of this thesis for form and content and recommend that it be accepted in partial fulfillment of the requirements for the degree of Master of Science, with a major in Computer Science.
    [Show full text]
  • AMD Core Math Library (ACML) ©
    ¨ AMD Core Math Library (ACML) © Version 6.1.0 Copyright c 2003-2014 Advanced Micro Devices, Inc., Numerical Algorithms Group Ltd. All rights reserved. The contents of this document are provided in connection with Advanced Micro Devices, Inc. (\AMD") products. AMD makes no representations or warranties with respect to the accuracy or completeness of the contents of this publication and reserves the right to make changes to specifications and product descriptions at any time without notice. The information contained herein may be of a preliminary or advance nature and is subject to change without notice. No license, whether express, implied, arising by estoppel, or oth- erwise, to any intellectual property rights are granted by this publication. Except as set forth in AMD's Standard Terms and Conditions of Sale, AMD assumes no liability whatso- ever, and disclaims any express or implied warranty, relating to its products including, but not limited to, the implied warranty of merchantability, fitness for a particular purpose, or infringement of any intellectual property right. AMD's products are not designed, intended, authorized or warranted for use as components in systems intended for surgical implant into the body, or in other applications intended to support or sustain life, or in any other application in which the failure of AMD's product could create a situation where personal injury, death, or severe property or environmental damage may occur. AMD reserves the right to discontinue or make changes to its products at any time without notice. Trademarks AMD, the AMD Arrow logo, and combinations thereof, AMD Athlon, and AMD Opteron are trademarks of Advanced Micro Devices, Inc.
    [Show full text]
  • Paket Som Matlab, Maple, Mathematica. • Generella
    Numerisk programvara NAG - The Numerical Algorithms Group Ltd Olika typer av programvara: http://www.nag.com/ • Paket som Matlab, Maple, Mathematica. NAG offers libraries in Fortran 77, Fortran 90, C, as well as • Generella rutinbibliotek, NAG, IMSL. F˚ar tillhandah˚alla an SMP Library and a Parallel Library for shared memory and huvudprogram etc. och l¨anka till en biblioteksfil. distributed memory parallel computing respectively. We have most recently added Data Mining Components to our product • Bibliotek f¨or speciella problemomr˚aden. range. Lapack (RISC-optimerat) rutiner f¨or Ax = b, minx ||Ax − b||2 och Ax = λx. Inneh˚aller s˚adar 1100 rutiner. (Linpack ¨aldre bibliotek f¨or CISC.) ¨ BLAS, t.ex. σ = xT y, y = Ax och C = AB. List of Contents, Release 4 Finns i NAG, IMSL men ¨aven i: Chapter 1: Utilities • Datorspecifika bibliotek, t.ex. Chapter 3: Special Functions – AMD: ACML (AMD Core Math Library) Chapter 4: Matrix and Vector Operations – Intel: MKL (Intel Math Kernel library). Chapter 5: Linear Equations – Sun: Sunperf (Sun Performance Library). – SGI: complib.sgimath Module 5.1: Solves a general real or complex system of linear (Scientific and Mathematical Library). equations with one or many right-hand sides – IBM: ESSL (Engineering and Scientific Subroutine Library). Module 5.2: Solves a real or complex, symmetric or Hermitian system of linear equations with one or many right-hand sides • Rutiner f¨or speciella till¨ampningar. Finns samlingar p˚awww. Netlib (d¨ar man ¨aven finner Lapack), Module 5.3: Solves a real or complex triangular system of linear Rutinerna ¨ar normalt skrivna i Fortran (finns Fortran66, 77, 90, equations 95).
    [Show full text]