Run-Time Power Modelling in Embedded Gpus with Dynamic Voltage and Frequency Scaling

Total Page:16

File Type:pdf, Size:1020Kb

Run-Time Power Modelling in Embedded Gpus with Dynamic Voltage and Frequency Scaling Run-Time Power Modelling in Embedded GPUs with Dynamic Voltage and Frequency Scaling Jose Nunez-Yanez, Kris Nikov, Kerstin Eder, Mohammad Hosseinabady {j.l.nunez-yanez,kris.nikov,kerstin.eder}@bristol.ac.uk,[email protected] Department of Electrical and Electronic Engineering, Bristol ABSTRACT popular as general-purpose accelerators in power constrained ap- This paper investigates the application of a robust CPU-based power plications such as unmanned aerial vehicles (UAV), self-driving cars modelling methodology that performs an automatic search of ex- or robotics. planatory events derived from performance counters to embedded The power profiles of these GPUs are in order of Watts compared GPUs. A 64-bit Tegra TX1 SoC is configured with DVFS enabled with the hundreds of Watts needed in their desktop counterparts and multiple CUDA benchmarks are used to train and test models that invariably use the PCIe bus to communicate with the host CPU optimized for each frequency and voltage point. These optimized and memory. Power optimization in these embedded devices is of models are then compared with a simpler unified model that uses a critical importance since in these applications, the power sources single set of model coefficients for all frequency and voltage points tend to be batteries and the systems must operate untethered for as of interest. To obtain this unified model, a number of experiments long as possible. In this paper we investigate the application of the are conducted to extract information on idle, clock and static power power modelling framework created in [6] for heterogeneous em- to derive power usage from a single reference equation. The re- bedded CPUs to embedded GPUs capable of general-purpose com- sults show that the unified model offers competitive accuracy with puting thanks to their support for languages such as as CUDA and an average 5% error with four explanatory variables on the test OpenCL. Our framework, called ROSE (RObust Statistical search of data set and it is capable to correctly predict the impact of voltage, explanatory activity Events), can be used to automatically collect frequency and temperature on power consumption. This model activity and power data and then perform a complete search for could be used to replace direct power measurements when these optimal events across a large range of frequency and voltage pairs are not available due to hardware limitations or worst-case analysis as defined in the device DVFS (Dynamic Voltage and Frequency in emulation platforms. Scaling) tables. ROSE multiple linear regression optimization uses OLS (Ordinary Least Squares) which is well understood for the case CCS CONCEPTS of both desktop and integrated CPUs but it has been less studied in GPUs that are characterized by a proprietary black box archi- • Hardware → Power estimation and optimization. tecture with restricted access to internal microarchitecture details. KEYWORDS Taking these points into account, the novelty of this work can be summarized as follows: keywords: Heterogeneous architecture, GPU power modelling, DVFS, multiple linear regression, Embedded GPU 1 We perform power modelling on an embedded GPU device ACM Reference Format: with integrated power measurement and voltage/frequency Jose Nunez-Yanez, Kris Nikov, Kerstin Eder, Mohammad Hosseinabady. 2020. scaling compared with previous work largely focused on Run-Time Power Modelling in Embedded GPUs with Dynamic Voltage and desktop GPUs. Frequency Scaling. In 11th Workshop on Parallel Programming and Run- 2 We limit the number of explanatory variables used in the Time Management Techniques for Many-core Architectures / 9th Workshop on model to enable the run-time collection of this information Design Tools and Architectures for Multicore Embedded Computing Platforms using a limited number of hardware registers. (PARMA-DITAM’20), January 21, 2020, Bologna, Italy. ACM, New York, NY, 3 We propose a novel unified model that includes temperature, arXiv:2006.12176v1 [cs.OH] 19 Jun 2020 USA, 6 pages. https://doi.org/10.1145/3381427.3381429 frequency and voltage as global states and compare it against models with coefficients optimized for single temperatures 1 INTRODUCTION and frequency/voltage pairs. Embedded GPUs (Graphical Processing Units) which are physically present in the same chip as the central processing unit (CPU) are This paper is organized as follows. Section 2 introduces related work in the area of power modelling with GPUs. Section 3 presents Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed the methodology based on our previous work in this area, the for profit or commercial advantage and that copies bear this notice and the full citation set of CUDA benchmarks for model training/verification and the on the first page. Copyrights for components of this work owned by others than the techniques used to obtain run-time measures of power and event author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission information. Section 4 develops models based on this methodology and/or a fee. Request permissions from [email protected]. with coefficients optimized for individual voltage and frequency PARMA-DITAM’20, January 21, 2020, Bologna, Italy pairs. Section 5 proposes an unified model so that the power ofa © 2020 Copyright held by the owner/author(s). Publication rights licensed to ACM. ACM ISBN 978-1-4503-7545-0/20/01...$15.00 single per-frequency model can be scaled to an extended range of https://doi.org/10.1145/3381427.3381429 voltage and frequency points. Section 6 investigates temperature PARMA-DITAM’20, January 21, 2020, Bologna, Italy Jose Nunez-Yanez, Kris Nikov, Kerstin Eder, Mohammad Hosseinabady effects on power consumption and model accuracy. Finally, section Training and testing benchmarks are independent and have been 7 concludes the paper. obtained from the Rodinia and CUDA SDK benchmark sets. Table 1: Deployed benchmarks 2 BACKGROUND CUDA Rodinia Train Set As previously indicated, there is a significant amount of work on stream_cluster srad_v1 srad_v1 power modelling of CPU cores and CPU-based systems and the particle_filter srad_v2 srad_v2 interested reader is referred to [7] for a review of results and tech- mmumergpu pathfinder pathfinder niques. In the field of general-purpose GPU-based computing, the leukocyte myocite myocite amount of power modelling research is more limited. The authors lavaMD kmeans kmeans of [4] investigate how performance counters can be used to model backprop bfs bfs power on a desktop NVIDIA GPUs connected to a host computer b+tree cfd cfd via a PCIe interface. The PCIe interface is instrumented with cur- heartwall hotspot3d hotspot3d rent clamp sensors and the host computer samples these sensors hotspot hybridsort hybridsort while collecting performance counter information. The authors CUDA SDK Test Set identify a total of 13 CUDA performance counters but since only binomialOptions Montecarlo four hardware registers are available, multiple runs are needed to blackscholes particles access all of them. The authors also identify that certain kernels that SobolQRNG Radixsort perform texture reads, such as Rodinia Leukocyte, show significant Transpose FDTD3d power error up to 50% due to lack of relevant counter information. Texture3D nbody The impact of DVFS on power modelling is not considered. Also targeting desktop GPUs, the authors of [8] introduce a support vec- tor regression model instead of the least squares linear regression We have modified the collection and processing stages to ac- more commonly used. A total of five variables are used, such as count for the differences in counter availability, power and current vertex shader busy and texture busy to build the model. Instead of sensors and DVFS implementations. In the CPU-based power model predicting the power of full kernels as done in [4], they predict done in [6] the DVFS table contains fixed pairs of voltage and fre- the power of the different execution phases as similarly done in quency and the preferred way to build the power model is to use a our work. The authors show a slight advantage of SVR in accuracy per-frequency model in which a distinct set of coefficient values are although some phases of execution of the GPU power cannot be calculated for each pair. To use this approach directly on the TX1 modelled correctly. The performance and power characterization is problematic due to temperature dependencies and the practical done in [1] considers different desktop NVIDIA GPU families (i.e. difficulties of adjusting the temperature of the device to each possi- Tesla, Fermi and Kepler). The external power measurements apply ble value during a data collection run that typically executes over to the entire system which includes the GPU and CPU and not the several days. To manage this complexity in this work we distinguish individual components. The proposed power model uses perfor- between local events such as the number of instructions executed mance counters and linear regression and introduces a frequency or the number of memory accesses that will affect power in cer- scaling parameter in the power equation to account for the different tain regions of the device, and global states such as the operating performance levels possible in the GPU. It does not consider the frequency, voltage and temperature that will affect power globally. operating voltage and, with multiple voltage levels possible for a This approach enables us to propose an unified model with a single single frequency, this could explain the errors in the prediction set of coefficients that could be used for multiple combinations accuracy which are measured at around 20 to 30%. The work of [3] of voltage, frequency and temperature. The development of this also focuses on desktop GPUs with a review that shows that the unified model and its comparison with the per-frequency models number of explanatory variables used varies between 8 and 23.
Recommended publications
  • GPU Architecture • Display Controller • Designing for Safety • Vision Processing
    i.MX 8 GRAPHICS ARCHITECTURE FTF-DES-N1940 RAFAL MALEWSKI HEAD OF GRAPHICS TECH ENGINEERING CENTER MAY 18, 2016 PUBLIC USE AGENDA • i.MX 8 Series Scalability • GPU Architecture • Display Controller • Designing for Safety • Vision Processing 1 PUBLIC USE #NXPFTF 1 PUBLIC USE #NXPFTF i.MX is… 2 PUBLIC USE #NXPFTF SoC Scalability = Investment Re-Use Replace the Chip a Increase capability. (Pin, Software, IP compatibility among parts) 3 PUBLIC USE #NXPFTF i.MX 8 Series GPU Cores • Dual Core GPU Up to 4 displays Accelerated ARM Cores • 16 Vec4 Shaders Vision Cortex-A53 | Cortex-A72 8 • Up to 128 GFLOPS • 64 execution units 8QuadMax 8 • Tessellation/Geometry total pixels Shaders Pin Compatibility Pin • Dual Core GPU Up to 4 displays Accelerated • 8 Vec4 Shaders Vision 4 • Up to 64 GFLOPS • 32 execution units 4 total pixels 8QuadPlus • Tessellation/Geometry Compatibility Software Shaders • Dual Core GPU Up to 4 displays Accelerated • 8 Vec4 Shaders Vision 4 • Up to 64 GFLOPS 4 • 32 execution units 8Quad • Tessellation/Geometry total pixels Shaders • Single Core GPU Up to 2 displays Accelerated • 8 Vec4 Shaders 2x 1080p total Vision • Up to 64 GFLOPS pixels Compatibility Pin 8 • 32 execution units 8Dual8Solo • Tessellation/Geometry Shaders • Single Core GPU Up to 2 displays Accelerated • 4 Vec4 Shaders 2x 1080p total Vision • Up to 32 GFLOPS pixels 4 • 16 execution units 8DualLite8Solo • Tessellation/Geometry Shaders 4 PUBLIC USE #NXPFTF i.MX 8 Series – The Doubles GPU Cores • Dual Core GPU Up to 4 displays Accelerated • 16 Vec4 Shaders Vision
    [Show full text]
  • Directx and GPU (Nvidia-Centric) History Why
    10/12/09 DirectX and GPU (Nvidia-centric) History DirectX 6 DirectX 7! ! DirectX 8! DirectX 9! DirectX 9.0c! Multitexturing! T&L ! SM 1.x! SM 2.0! SM 3.0! DirectX 5! Riva TNT GeForce 256 ! ! GeForce3! GeForceFX! GeForce 6! Riva 128! (NV4) (NV10) (NV20) Cg! (NV30) (NV40) XNA and Programmable Shaders DirectX 2! 1996! 1998! 1999! 2000! 2001! 2002! 2003! 2004! DirectX 10! Prof. Hsien-Hsin Sean Lee SM 4.0! GTX200 NVidia’s Unified Shader Model! Dualx1.4 billion GeForce 8 School of Electrical and Computer 3dfx’s response 3dfx demise ! Transistors first to Voodoo2 (G80) ~1.2GHz Engineering Voodoo chip GeForce 9! Georgia Institute of Technology 2006! 2008! GT 200 2009! ! GT 300! Adapted from David Kirk’s slide Why Programmable Shaders Evolution of Graphics Processing Units • Hardwired pipeline • Pre-GPU – Video controller – Produces limited effects – Dumb frame buffer – Effects look the same • First generation GPU – PCI bus – Gamers want unique look-n-feel – Rasterization done on GPU – Multi-texturing somewhat alleviates this, but not enough – ATI Rage, Nvidia TNT2, 3dfx Voodoo3 (‘96) • Second generation GPU – Less interoperable, less portable – AGP – Include T&L into GPU (D3D 7.0) • Programmable Shaders – Nvidia GeForce 256 (NV10), ATI Radeon 7500, S3 Savage3D (’98) – Vertex Shader • Third generation GPU – Programmable vertex and fragment shaders (D3D 8.0, SM1.0) – Pixel or Fragment Shader – Nvidia GeForce3, ATI Radeon 8500, Microsoft Xbox (’01) – Starting from DX 8.0 (assembly) • Fourth generation GPU – Programmable vertex and fragment shaders
    [Show full text]
  • MSI Afterburner V4.6.4
    MSI Afterburner v4.6.4 MSI Afterburner is ultimate graphics card utility, co-developed by MSI and RivaTuner teams. Please visit https://msi.com/page/afterburner to get more information about the product and download new versions SYSTEM REQUIREMENTS: ...................................................................................................................................... 3 FEATURES: ............................................................................................................................................................. 3 KNOWN LIMITATIONS:........................................................................................................................................... 4 REVISION HISTORY: ................................................................................................................................................ 5 VERSION 4.6.4 .............................................................................................................................................................. 5 VERSION 4.6.3 (PUBLISHED ON 03.03.2021) .................................................................................................................... 5 VERSION 4.6.2 (PUBLISHED ON 29.10.2019) .................................................................................................................... 6 VERSION 4.6.1 (PUBLISHED ON 21.04.2019) .................................................................................................................... 7 VERSION 4.6.0 (PUBLISHED ON
    [Show full text]
  • Graphics Shaders Mike Hergaarden January 2011, VU Amsterdam
    Graphics shaders Mike Hergaarden January 2011, VU Amsterdam [Source: http://unity3d.com/support/documentation/Manual/Materials.html] Abstract Shaders are the key to finalizing any image rendering, be it computer games, images or movies. Shaders have greatly improved the output of computer generated media; from photo-realistic computer generated humans to more vivid cartoon movies. Furthermore shaders paved the way to general-purpose computation on GPU (GPGPU). This paper introduces shaders by discussing the theory and practice. Introduction A shader is a piece of code that is executed on the Graphics Processing Unit (GPU), usually found on a graphics card, to manipulate an image before it is drawn to the screen. Shaders allow for various kinds of rendering effect, ranging from adding an X-Ray view to adding cartoony outlines to rendering output. The history of shaders starts at LucasFilm in the early 1980’s. LucasFilm hired graphics programmers to computerize the special effects industry [1]. This proved a success for the film/ rendering industry, especially at Pixars Toy Story movie launch in 1995. RenderMan introduced the notion of Shaders; “The Renderman Shading Language allows material definitions of surfaces to be described in not only a simple manner, but also highly complex and custom manner using a C like language. Using this method as opposed to a pre-defined set of materials allows for complex procedural textures, new shading models and programmable lighting. Another thing that sets the renderers based on the RISpec apart from many other renderers, is the ability to output arbitrary variables as an image—surface normals, separate lighting passes and pretty much anything else can be output from the renderer in one pass.” [1] The term shader was first only used to refer to “pixel shaders”, but soon enough new uses of shaders such as vertex and geometry shaders were introduced, making the term shaders more general.
    [Show full text]
  • CPU-GPU Benchmark Description
    UC San Diego UC San Diego Electronic Theses and Dissertations Title Heterogeneity and Density Aware Design of Computing Systems Permalink https://escholarship.org/uc/item/9qd6w8cw Author Arora, Manish Publication Date 2018 Peer reviewed|Thesis/dissertation eScholarship.org Powered by the California Digital Library University of California UNIVERSITY OF CALIFORNIA, SAN DIEGO Heterogeneity and Density Aware Design of Computing Systems A dissertation submitted in partial satisfaction of the requirements for the degree Doctor of Philosophy in Computer Science (Computer Engineering) by Manish Arora Committee in charge: Professor Dean M. Tullsen, Chair Professor Renkun Chen Professor George Porter Professor Tajana Simunic Rosing Professor Steve Swanson 2018 Copyright Manish Arora, 2018 All rights reserved. The dissertation of Manish Arora is approved, and it is ac- ceptable in quality and form for publication on microfilm and electronically: Chair University of California, San Diego 2018 iii DEDICATION To my family and friends. iv EPIGRAPH Do not be embarrassed by your failures, learn from them and start again. —Richard Branson If something’s important enough, you should try. Even if the probable outcome is failure. —Elon Musk v TABLE OF CONTENTS Signature Page . iii Dedication . iv Epigraph . v Table of Contents . vi List of Figures . viii List of Tables . x Acknowledgements . xi Vita . xiii Abstract of the Dissertation . xvi Chapter 1 Introduction . 1 1.1 Design for Heterogeneity . 2 1.1.1 Heterogeneity Aware Design . 3 1.2 Design for Density . 6 1.2.1 Density Aware Design . 7 1.3 Overview of Dissertation . 8 Chapter 2 Background . 10 2.1 GPGPU Computing . 10 2.2 CPU-GPU Systems .
    [Show full text]
  • Model System of Zirconium Oxide An
    First-principles based treatment of charged species equilibria at electrochemical interfaces: model system of zirconium oxide and titanium oxide by Jing Yang B.S., Physics, Peking University, China Submitted to the Department of Materials Science and Engineering in partial fulfillment of the requirements for the degree of Doctor of Philosophy at the MASSACHUSETTS INSTITUTE OF TECHONOLOGY February 2020 © Massachusetts Institute of Technology 2020. All rights reserved. Signature of Author: Jing Yang Department of Materials Science and Engineering January 10th, 2020 Certified by: Bilge Yildiz Professor of Materials Science and Engineering And Professor of Nuclear Science and Engineering Thesis Supervisor Accepted by: Donald R. Sadoway John F. Elliott Professor of Materials Chemistry Chair, Department Committee on Graduate Studies 2 First-principles based treatment of charged species equilibria at electrochemical interfaces: model system of zirconium oxide and titanium oxide by Jing Yang Submitted to the Department of Materials Science and Engineering on Jan. 10th, 2020 in partial fulfillment of the requirements for the degree of Doctor of Philosophy Abstract Ionic defects are known to influence the magnetic, electronic and transport properties of oxide materials. Understanding defect chemistry provides guidelines for defect engineering in various applications including energy storage and conversion, corrosion, and neuromorphic computing devices. While DFT calculations have been proven as a powerful tool for modeling point defects in bulk oxide materials, challenges remain for linking atomistic results with experimentally-measurable materials properties, where impurities and complicated microstructures exist and the materials deviate from ideal bulk behavior. This thesis aims to bridge this gap on two aspects. First, we study the coupled electro-chemo-mechanical effects of ionic impurities in bulk oxide materials.
    [Show full text]
  • ATI Radeon Driver for Plan 9 Implementing R600 Support
    Introduction Radeon Specifics Plan 9 Specifics Conclusion ATI Radeon Driver for Plan 9 Implementing R600 Support Sean Stangl ([email protected]) 15-412, Carnegie Mellon University October 23, 2009 1 Introduction Radeon Specifics Plan 9 Specifics Conclusion Introduction Chipset Schema Project Goal Radeon Specifics Available Documentation ATOMBIOS Command Processor Plan 9 Specifics Structure of Plan 9 Graphics Drivers Conclusion Figures and Estimates Questions 2 Introduction Radeon Specifics Plan 9 Specifics Conclusion Chipset Schema Radeon Numbering System I ATI has released many cards under the Radeon label. I Modern cards are prefixed with "HD". Rough Mapping from Chipset to Marketing Name R200 : Radeon {8500, 9000, 9200, 9250} R300 : Radeon {9500, 9600, 9700, 9800} R420 : Radeon {X700, X740, X800, X850} R520 : Radeon {X1300, X1600, X1800, X1900} R600 : Radeon HD {2400, 3600, 3800, 3870 X2} R700 : Radeon HD {4300, 4600, 4800, 4800 X2} 3 I R520 introduces yet another new shader model. I ATI develops ATOMBIOS, a collection of data tables and scripts stored in the ROM on each card. I R600 supports the Unified Shader Model. I The 2D engine gets unified into the 3D engine. I The register specification is completely different. I R700 is an optimized R600. Introduction Radeon Specifics Plan 9 Specifics Conclusion Chipset Schema Notable Chipset Deltas Each family roughly coincides with a DirectX version bump. I R200, R300, and R420 use a similar architecture. I Each family changes the pixel shader. 4 I R600 supports the Unified Shader Model. I The 2D engine gets unified into the 3D engine. I The register specification is completely different.
    [Show full text]
  • 04-Prog-On-Gpu-Schaber.Pdf
    Seminar - Paralleles Rechnen auf Grafikkarten Seminarvortrag: Übersicht über die Programmierung von Grafikkarten Marcus Schaber 05.05.2009 Betreuer: J. Kunkel, O. Mordvinova 05.05.2009 Marcus Schaber 1/33 Gliederung ● Allgemeine Programmierung Parallelität, Einschränkungen, Vorteile ● Grafikpipeline Programmierbare Einheiten, Hardwarefunktionalität ● Programmiersprachen Übersicht und Vergleich, Datenstrukturen, Programmieransätze ● Shader Shader Standards ● Beispiele Aus Bereich Grafik 05.05.2009 Marcus Schaber 2/33 Programmierung Einschränkungen ● Anpassung an Hardware Kernels und Streams müssen erstellt werden. Daten als „Vertizes“, Erstellung geeigneter Fragmente notwendig. ● Rahmenprogramm Programme können nicht direkt von der GPU ausgeführt werden. Stream und Kernel werden von der CPU erstellt und auf die Grafikkarte gebracht 05.05.2009 Marcus Schaber 3/33 Programmierung Parallelität ● „Sequentielle“ Sprachen Keine spezielle Syntax für parallele Programmierung. Einzelne shader sind sequentielle Programme. ● Parallelität wird durch gleichzeitiges ausführen eines Shaders auf unterschiedlichen Daten erreicht ● Hardwareunterstützung Verwaltung der Threads wird von Hardware unterstützt 05.05.2009 Marcus Schaber 4/33 Programmierung Typische Aufgaben ● Geeignet: Datenparallele Probleme Gleiche/ähnliche Operationen auf vielen Daten ausführen, wenig Kommunikation zwischen Elementen notwendig ● Arithmetische Dichte: Anzahl Rechenoperationen / Lese- und Schreibzugiffe ● Klassisch: Grafik Viele unabhängige Daten: Vertizes, Fragmente/Pixel Operationen:
    [Show full text]
  • 1 Títol: Aceleración Con CUDA De Procesos 3D Volum
    84 mm Títol: Aceleración con CUDA de Procesos 3D Volum: 1/1 Alumne: Jose Antonio Raya Vaquera Director/Ponent: Enric Xavier Martin Rull 55 mm Departament: ESAII Data: 26/01/2009 1 2 DADES DEL PROJECTE Títol del Projecte: Aceleración con CUDA de procesos 3D Nom de l'estudiant: Jose Antonio Raya Vaquera Titulació: Enginyeria en Informàtica Crèdits: 37.5 Director/Ponent: Enric Xavier Martín Rull Departament: ESAII MEMBRES DEL TRIBUNAL (nom i signatura) President: Vocal: Secretari: QUALIFICACIÓ Qualificació numèrica: Qualificació descriptiva: Data: 3 4 ÍNDICE Índice de figuras ............................................................................................................... 9 Índice de tablas ............................................................................................................... 12 1. Introducción ............................................................................................................ 13 1.1 Motivación ....................................................................................................... 13 1.2 Objetivos .......................................................................................................... 14 2. Antecedentes ........................................................................................................... 17 3. Historia .................................................................................................................... 21 3.1 Evolución de las tarjetas gráficas ....................................................................
    [Show full text]
  • HP Z800 Workstation Overview
    QuickSpecs HP Z800 Workstation Overview HP recommends Windows Vista® Business 1. 3 External 5.25" Bays 2. Power Button 3. Front I/O: 3 USB 2.0, 1 IEEE 1394a, 1 Headphone out, 1 Microphone in DA - 13278 North America — Version 2 — April 13, 2009 Page 1 QuickSpecs HP Z800 Workstation Overview 4. Choice of 850W, 85% or 1110W, 89% Power Supplies 9. Rear I/O: 1 IEEE 1394a, 6 USB 2.0, 1 serial, PS/2 keyboard/mouse 5. 12 DIMM Slots for DDR3 ECC Memory 2 RJ-45 to Integrated Gigabit LAN 1 Audio Line In, 1 Audio Line Out, 1 Microphone In 6. 3 External 5.25” Bays 10. 2 PCIe x16 Gen2 Slots 7. 4 Internal 3.5” Bays 11.. 2 PCIe x8 Gen2, 1 PCIe x4 Gen2, 1 PCIe x4 Gen1, 1 PCI Slot 8. 2 Quad Core Intel 5500 Series Processors 12 3 Internal USB 2.0 ports Form Factor Rackable Minitower Compatible Operating Genuine Windows Vista® Business 32-bit* Systems Genuine Windows Vista® Business 64-bit* Genuine Windows Vista® Business 32-bit with downgrade to Windows® XP Professional 32-bit custom installed** Genuine Windows Vista® Business 64-bit with downgrade to Windows® XP Professional x64 custom installed** HP Linux Installer Kit for Linux (includes drivers for both 32-bit & 64-bit OS versions of Red Hat Enterprise Linux WS4 and WS5 - see: http://www.hp.com/workstations/software/linux) For detailed OS/hardware support information for Linux, see: http://www.hp.com/support/linux_hardware_matrix *Certain Windows Vista product features require advanced or additional hardware.
    [Show full text]
  • Nvidia's GPU Microarchitectures
    NVidia’s GPU Microarchitectures By Stephen Lucas and Gerald Kotas Intro Discussion Points - Difference between CPU and GPU - Use s of GPUS - Brie f History - Te sla Archite cture - Fermi Architecture - Kepler Architecture - Ma xwe ll Archite cture - Future Advances CPU vs GPU Architectures ● Few Cores ● Hundreds of Cores ● Lots of Cache ● Thousands of Threads ● Handful of Threads ● Single Proce ss Exe cution ● Independent Processes Use s of GPUs Gaming ● Gra phics ● Focuses on High Frames per Second ● Low Polygon Count ● Predefined Textures Workstation ● Computation ● Focuses on Floating Point Precision ● CAD Gra p hics - Billions of Polygons Brief Timeline of NVidia GPUs 1999 - World’s First GPU: GeForce 256 2001 - First Programmable GPU: GeForce3 2004 - Sca la ble Link Inte rfa ce 2006 - CUDA Architecture Announced 2007 - Launch of Tesla Computation GPUs with Tesla Microarchitecture 2009 - Fermi Architecture Introduced 2 0 13 - Kepler Architecture Launched 2 0 16 - Pascal Architecture Tesla - First microarchitecture to implement the unified shader model, which uses the same hardware resources for all fragment processing - Consists of a number of stream processors, which are scalar and can only operate on one component at a time - Increased clock speed in GPUs - Round robin scheduling for warps - Contained local, shared, and global memory - Conta ins Spe cia l-Function Units which are specialized for interpolating points - Allowed for two instructions to execute per clock cycle per SP Fermi Peak Performance Overview: ● Each SM has 32
    [Show full text]
  • Evaluating ATTILA, a Cycle-Accurate GPU Simulator
    Evaluating ATTILA, a cycle-accurate GPU simulator Miguel Ángel Martínez del Amor Department of Computer and Information Science Norwegian University of Science and Technology (NTNU) [email protected] Project supervisor: Lasse Natvig [email protected] 26 th January 2007 Abstract. As GPUs’ technology grow more and more, new simulators are needed to help on the task of designing new architectures. Simulators have several advantages compared to an implementation directly in hardware, so they are very useful in GPU development. ATTILA is a new GPU simulator that was born to provide a framework for working with real OpenGL applications, simulating at the same time a GPU that has an architecture similar to the current designs of the major GPU vendors like NVIDIA or ATI. This architecture is highly configurable, and also ATTILA gives a lot of statistics and information about what occurred into the simulated GPU during the execution. This project will explain the main aspects of this simulator, with an evaluation of this tool that is young but powerful and helpful for researchers and developers of new GPUs. 1. Introduction The main brain of computers is the CPU (Central Processing Unit), where the instructions are executed, manipulating by this way the data stored in memory. But with the new applications that demand even more resources, the CPU is becoming a bottleneck. There is a lot of works about how to improve the CPI (Cycles per Instruction) which is a good parameter for making comparisons between uniprocessor architectures, and the search for parallelism is the best way to do it.
    [Show full text]