Phd Thesis, Norwegian University of Science and Technology, 2013

Total Page:16

File Type:pdf, Size:1020Kb

Phd Thesis, Norwegian University of Science and Technology, 2013 i “phdthesis” — 2015/12/14 — 9:35 — page 1 — #1 i i i POLITECNICO DI MILANO DEPARTMENT OF INFORMATION TECHNOLOGY DOCTORAL PROGRAMME IN COMPUTER SCIENCE AND ENGINEERING ON THE ROLE OF POLYHEDRAL ANALYSIS IN HIGH PERFORMANCE RECONFIGURABLE HARDWARE BASED COMPUTING SYSTEMS Doctoral Dissertation of: Riccardo Cattaneo Advisor: Prof. Marco D. Santambrogio Tutor: Prof. Donatella Sciuto The Chair of the Doctoral Program: Prof. Carlo Fiorini i i i i i “phdthesis” — 2015/12/14 — 9:35 — page 2 — #2 i i i 2015 – XVIII 2 i i i i i “phdthesis” — 2015/12/14 — 9:35 — page 1 — #3 i i i Acknowledgements 1 i i i i i “phdthesis” — 2015/12/14 — 9:35 — page 2 — #4 i i i i i i i i “phdthesis” — 2015/12/14 — 9:35 — page I — #5 i i i Summary ECENT years have seen dramatic improvements in High Level Synthesis (HLS) tools: from efficient translation of numeric algorithms to image processing algorithms, from stencil com- Rputations to neural networks, relevant application domains and in- dustries are benefiting from research on compute and/or memory/- communications network. In order to systematically synthesize better circuits for specific programs and kernels, last decades’ studies focused on the develop- ment of sound, formal approaches; one notable such framework is the Polyhedral Model (PM) and the associated code analysis tech- nique, collectively called Polyhedral Analysis (PA). Under this rep- resentation it is possible to compute dependencies, find loop bounds, and reorder instructions in a completely automated manner relying on the same set of sound and comprehensive assumptions of PM. We reconsider this rich state of the art, and elaborate on differ- ent methodologies and implement the related toolchains to gener- ate highly parallel and power efficient kernels running on recon- figurable hardware, with the aim of distributing the workload on multiple computational blocks while maximizing the overall power efficiency of the resulting heterogeneous system. We approach this problem in three different, interrelated ways. First of all, we dedicated our early efforts to the development of an accelerator-rich platform where the focus is on the coordina- tion of multiple custom and software processors, via a novel Do- I i i i i i “phdthesis” — 2015/12/14 — 9:35 — page II — #6 i i i main Space Exploration (DSE) phase. Specifically, the work elabo- rates on two relevant aspects: the effectiveness of Partial Reconfigu- ration (PR) to attain improved energy delay and throughput metrics, and the effectiveness of the heuristics chosen to realize the DSE step, which feature both low complexity and good exploration times. To this matter, we devote Chapter 2. We also extended the scope of this work by assuming that multiple computing elements are in place, and an adequate communication architecture is required to coordi- nate those accelerators. This is discussed in Chapter 3. Secondly, we focus on amply data-parallel codes, and develop a novel HLS approach to using PM as a means to explicitly extract and isolate data and computation from affine codes in order to effi- ciently divide the workload among an arbitrary number of nodes, in the light of the current and foreseeable trend of adoption of recon- figurable hardware in the datacenter; towards energy proportional computing, we improve the current state of art in single core accel- eration, as our methodology obtains near-linear speedup with the area at disposition to accelerate the given workload. To this subject, we devote Chapter 4 and 5. Lastly, we focus on a specific, and more restricted class of data parallel codes, namely Iterative Stencil Loop (ISL), as they play a crucial role in a variety of different fields of application. The com- putationally intensive nature of those algorithms created the need for solutions to efficiently implement them in order to save both ex- ecution time and energy. This, in combination with their regular structure, has justified their widespread study and the proposal of largely different approaches to their optimization. However, most of these works are focused on aggressive compile time optimization, cache locality optimization, and parallelism extraction for the mul- ticore/multi processor domain, while fewer works are focused on the exploitation of custom architectures to further exploit the regular structure of ISLs, specifically with the goal of improving power ef- ficiency. This work introduces a methodology to systematically de- sign power efficient hardware accelerators for the optimal execution of ISLs algorithms on Field Programmable Gate Arrays (FPGAs). To this extensive methodology, we devote Chapter 6. As part of the methodology, we introduce the notion of Stream- ing Stencil Time-step (SST), a streaming-based architecture capable of achieving both low resource usage and efficient data reuse thanks to an optimal data buffering strategy; and we introduce a technique, II i i i i i “phdthesis” — 2015/12/14 — 9:35 — page III — #7 i i i called SSTs queuing, capable to deliver a quasi-linear execution time speedup with constant bandwidth. This is the core subject of Chap- ter 7. We validate all methodologies, approaches, and toolchains on significant benchmarks on either a Zynq-7000 or Virtex-7 FPGA us- ing the Xilinx Vivado suite. Results, which are reported in each Chap- ter devoted to the respective subject, demonstrate how we are able to improve the efficiency of all the baselines we compare against; specifically, we improve the energy delay and throughput metric when using our accelerator-rich platform, and dramatically improve the on-chip memory resources usage using the PA-based method- ologies, allowing us to treat problem sizes whose implementation would otherwise not be possible via direct synthesis of the original, unmanipulated HLS code. We finally conclude the dissertation in Chapter 8 with a number of potential future works. III i i i i i “phdthesis” — 2015/12/14 — 9:35 — page IV — #8 i i i i i i i i “phdthesis” — 2015/12/14 — 9:35 — page V — #9 i i i Sommario URANTE gli ultimi anni gli strumenti di sintesi ad alto liv- ello (High Level Synthesis (HLS)) hanno conosciuto notevoli miglioramenti: dalla traduzione efficiente di algoritmi nu- Dmerici a quella degli algoritmi di elaborazione delle immagini, dai calcoli stencil alle reti neurali, disparati domini applicativi ed indus- trie stanno beneficiando della ricerca sulla sintesi di elementi di cal- colo, sottosistemi di memoria, e reti di comunicazione. La ricerca degli ultimi decenni si é focalizzata sullo sviluppo di approcci formali e corretti finalizzati alla sistematica sintesi di cir- cuiti piú performanti per specifici programmi e nuclei computazion- ali complessi. In tale senso é di notevole interesse il Modello Poliedrale (Polyhedral Model (PM)) e la tecnica di analisi codice associata, cui ci si riferisce collettivamente con il termine Analisi Poliedrale (Poly- hedral Analysis (PA)), descritta con dovizia di particolari nel capi- tolo 4. Tramite questa rappresentazione e manipolazione della com- putazione é possibile calcolare le dipendenze, i gli indici degli es- tremi delle iterazioni, riordinare le istruzioni in modo completamente automatizzato, ed altro ancora. In questa tesi riconsideriamo lo stato dell’arte ed i lavori relativi a quest’area di ricerca, elaborando metodolo- gie innovative ed implementando le relative toolchain per generare hardware riconfigurabile, con l’obiettivo di distribuire il carico di lavoro su piú blocchi di calcolo, massimizzando l’efficienza energet- ica complessiva di il sistema eterogeneo risultante. Approcciamo il problema in tre modi differenti. V i i i i i “phdthesis” — 2015/12/14 — 9:35 — page VI — #10 i i i Prima di tutto, abbiamo sviluppato una piattaforma basata su acceleratori in cui l’attenzione é rivolta al coordinamento di piú pro- cessori e software personalizzati, attraverso una fase di esplorazione dello spazio dei parametri (Domain Space Exploration (DSE)). Il la- voro approfondisce due aspetti rilevanti: primo, l’efficacia della ri- configurazione parziale (Partial Reconfiguration (PR)) per ottenere un miglioramento delle metriche prodotto energia-latenza e through- put; secondo, l’efficacia delle euristiche scelte per realizzare il passo DSE, di bassa complessitá computazionale e con ottimi tempi di es- ecuzione. A questo lavoro dedichiamo il capitolo 2. Abbiamo anche ampliato la portata dello stesso assumendo che piú elementi di cal- colo siano presenti nel sistema, e sia necessario un’adeguata architet- tura di comunicazione al fine di coordinare tali acceleratori. Questo é discusso nel capitolo 3. Successivamente ci concentriamo sui codici affini ampiamente paralleli, e sviluppiamo un nuovo approccio alla sintesi ad alto liv- ello (HLS) per utilizzare l’analisi poliedrale come un mezzo per es- trarre ed isolare flusso dati e computazione, per poter dividere in modo efficiente il carico di lavoro tra un numero arbitrario di nodi alla luce della tendenza attuale e futura di adozione di hardware ri- configurabile nei centri di elaborazione dati; nell’ottica dei sistemi di calcolo energy proportional, miglioriamo l’attuale stato dell’arte in termini di accelerazione single core, come dimostra la metodologia descritta nel capitolo 5. Infine, ci concentriamo su una classe specifica e ristretta di cod- ici paralleli, gli iterative Stencil Loop (ISL), codici la cui computazione efficiente risulta cruciale in una varietá di campi di applicazione. La natura computazionalmente intensiva di questi algoritmi ha gen- erato la necessitá di disporre di metodi efficienti per la loro com- putazione al fine di risparmiare sia tempo che energia per la loro es- ecuzione. Questo, in combinazione con la loro struttura regolare, ha giustificato il loro ampio studio e la proposta di molteplici approcci per loro ottimizzazione. Tuttavia, la maggior parte di queste opere sono focalizzate sull’ottimizzazione statica aggressiva, l’ottimizzazione della localitá dei dati, e l’estrazione di parallelismo nel dominio mul- ticore, mentre un minor numero di lavori sono focalizzati sullo sfrut- tamento di architetture riconfigurabili con l’obiettivo di migliorare l’efficienza energetica della computazione.
Recommended publications
  • CUDA Particle-Based Fluid Simulation
    CUDA Particle-based Fluid Simulation Simon Green Overview Fluid Simulation Techniques CUDA particle simulation Spatial subdivision techniques Rendering methods Future © NVIDIA Corporation 2008 Fluid Simulation Techniques Various approaches: Grid based (Eulerian) Stable fluids Particle level set Particle based (Lagrangian) SPH (smoothed particle hydrodynamics) MPS (Moving-Particle Semi-Implicit) Height field FFT (Tessendorf) Wave propagation – e.g. Kass and Miller © NVIDIA Corporation 2008 CUDA N-Body Demo Computes gravitational attraction between n bodies Computes all n2 interactions Uses shared memory to reduce memory bandwidth 16K bodies @ 44 FPS x 20 FLOPS / interaction x 16K2 interactions / frame = 240 GFLOP/s GeForce 8800 GTX © NVIDIA Corporation 2008 Particle-based Fluid Simulation Advantages Conservation of mass is trivial Easy to track free surface Only performs computation where necessary Not necessarily constrained to a finite grid Easy to parallelize Disadvantages Hard to extract smooth surface from particles Requires large number of particles for realistic results © NVIDIA Corporation 2008 Particle Fluid Simulation Papers Particle-Based Fluid Simulation for Interactive Applications, M. Müller, 2003 3000 particles, 5fps Particle-based Viscoelastic Fluid Simulation, Clavet et al, 2005 1000 particles, 10fps 20,000 particles, 2 secs / frame © NVIDIA Corporation 2008 CUDA SDK Particles Demo Particles with simple collisions Uses uniform grid based on sorting Uses fast CUDA radix sort Current performance: >100 fps for 65K interacting
    [Show full text]
  • NVIDIA Physx SDK EULA
    NVIDIA CORPORATION NVIDIA® PHYSX® SDK END USER LICENSE AGREEMENT Welcome to the new world of reality gaming brought to you by PhysX® acceleration from NVIDIA®. NVIDIA Corporation (“NVIDIA”) is willing to license the PHYSX SDK and the accompanying documentation to you only on the condition that you accept all the terms in this License Agreement (“Agreement”). IMPORTANT: READ THE FOLLOWING TERMS AND CONDITIONS BEFORE USING THE ACCOMPANYING NVIDIA PHYSX SDK. IF YOU DO NOT AGREE TO THE TERMS OF THIS AGREEMENT, NVIDIA IS NOT WILLING TO LICENSE THE PHYSX SDK TO YOU. IF YOU DO NOT AGREE TO THESE TERMS, YOU SHALL DESTROY THIS ENTIRE PRODUCT AND PROVIDE EMAIL VERIFICATION TO [email protected] OF DELETION OF ALL COPIES OF THE ENTIRE PRODUCT. NVIDIA MAY MODIFY THE TERMS OF THIS AGREEMENT FROM TIME TO TIME. ANY USE OF THE PHYSX SDK WILL BE SUBJECT TO SUCH UPDATED TERMS. A CURRENT VERSION OF THIS AGREEMENT IS POSTED ON NVIDIA’S DEVELOPER WEBSITE: www.developer.nvidia.com/object/physx_eula.html 1. Definitions. “Licensed Platforms” means the following: - Any PC or Apple Mac computer with a NVIDIA CUDA-enabled processor executing NVIDIA PhysX; - Any PC or Apple Mac computer running NVIDIA PhysX software executing on the primary central processing unit of the PC only; - Any PC utilizing an AGEIA PhysX processor executing NVIDIA PhysX code; - Microsoft XBOX 360; - Nintendo Wii; and/or - Sony Playstation 3 “Physics Application” means a software application designed for use and fully compatible with the PhysX SDK and/or NVIDIA Graphics processor products, including but not limited to, a video game, visual simulation, movie, or other product.
    [Show full text]
  • Multiprocessing Contents
    Multiprocessing Contents 1 Multiprocessing 1 1.1 Pre-history .............................................. 1 1.2 Key topics ............................................... 1 1.2.1 Processor symmetry ...................................... 1 1.2.2 Instruction and data streams ................................. 1 1.2.3 Processor coupling ...................................... 2 1.2.4 Multiprocessor Communication Architecture ......................... 2 1.3 Flynn’s taxonomy ........................................... 2 1.3.1 SISD multiprocessing ..................................... 2 1.3.2 SIMD multiprocessing .................................... 2 1.3.3 MISD multiprocessing .................................... 3 1.3.4 MIMD multiprocessing .................................... 3 1.4 See also ................................................ 3 1.5 References ............................................... 3 2 Computer multitasking 5 2.1 Multiprogramming .......................................... 5 2.2 Cooperative multitasking ....................................... 6 2.3 Preemptive multitasking ....................................... 6 2.4 Real time ............................................... 7 2.5 Multithreading ............................................ 7 2.6 Memory protection .......................................... 7 2.7 Memory swapping .......................................... 7 2.8 Programming ............................................. 7 2.9 See also ................................................ 8 2.10 References .............................................
    [Show full text]
  • Nvidia Corporation 2016 Annual Report
    2017 NVIDIA CORPORATION ANNUAL REVIEW NOTICE OF ANNUAL MEETING PROXY STATEMENT FORM 10-K THE AGE OF THE GPU IS UPON US THE NEXT PLATFORM A decade ago, we set out to transform the GPU into a powerful computing platform—a specialized tool for the da Vincis and Einsteins of our time. GPU computing has since opened a floodgate of innovation. From gaming and VR to AI and self-driving cars, we’re at the center of the most promising trends in our lifetime. The world has taken notice. Jensen Huang CEO and Founder, NVIDIA NVIDIA GEFORCE HAS MOVED FROM GRAPHICS CARD TO GAMING PLATFORM FORBES The PC drives the growth of computer gaming, the largest entertainment industry in the world. The mass popularity of eSports, the evolution of gaming into a social medium, and the advance of new technologies like 4K, HDR, and VR will fuel its further growth. Today, 200 million gamers play on GeForce, the world’s largest gaming platform. Our breakthrough NVIDIA Pascal architecture delighted gamers, and we built on its foundation with new capabilities, like NVIDIA Ansel, the most advanced in-game photography system ever built. To serve the 1 billion new or infrequent gamers whose PCs are not ready for today’s games, we brought the GeForce NOW game streaming service to Macs and PCs. Mass Effect: Andromeda. Courtesy of Electronic Arts. THE BEST ANDROID TV DEVICE JUST GOT BETTER ENGADGET NVIDIA SHIELD TV, controller, and remote. NVIDIA SHIELD is taking NVIDIA gaming and AI into living rooms around the world. The most advanced streamer now boasts 1,000 games, has the largest open catalog of media in 4K, and can serve as the brain of the AI home.
    [Show full text]
  • AMD Pcie® ADD-IN BOARD
    NVIDIA GT610 PCIe® for passive ADD-IN BOARD Datasheet AEGX-N3A1-01FSS1 CONTENTS 1. Feature .................................................................................................................. 3 2. Functional Overview ............................................................................................. 4 2.1. GPU Block diagram .................................................................................... 4 2.2. GPU ............................................................................................................ 4 2.3. Memory Interface ..................................................................................... 5 2.4. Features and Technologies ....................................................................... 5 2.5. Display Support ......................................................................................... 5 2.6. Digital Audio .............................................................................................. 5 2.7. Video ......................................................................................................... 6 3. PIN Assignment and Description………………………………………………………………………..7 3.1. DVI-I Connector Pinout ............................................................................ .7 3.2. HDMI 1.4a Connector Pinout ................................................................... .8 3.3. VGA Connector Pinout ............................................................................. .9 3.4. VGA Header Pinout ................................................................................
    [Show full text]
  • Inside Chips
    InsideChips.VenturesTM Tracking Fabless, IP & Design-House Startups Volume 6, Number 5 May 2005 Business Microscope Semiconductor Forecast ... Each year, analyzing capital spending and fab capacity trends, InsideChips bestows a Semiconductor Forecast researching ASIC markets and technologies, and Award to the best analyst forecaster after actual following emerging markets for ICs such as cellular results are posted by the SIA. The purpose of the phones. Feldhan is the founder of SEMICO, one of program is to identify and recognize top analysts the several market research boutiques that spun out whose methodology and experience result in of In-Stat during the 1980s. SEMICO is a consistently accurate forecasting. semiconductor-centric market research firm whose Based on forecasts made in Q4 2003, the award services include consulting, reports, events and for the best semiconductor forecast made by a market newsletters. research firm for 2004 is shared by Bill McClean of The most conservative forecasts for 2004 were IC Insights and Jim Feldhan of SEMICO Research. Both analysts 17%, 18% and 19% made by iSupply, IDC, and SIA, respectively. forecasted +27% growth for 2004, which came in very close to VLSI Research was most optimistic at 32%. The semiconductor the actual +28% growth over 2003 as announced by the forecasts for 2004 average 25.6%. Semiconductor Industry Association. McClean also won This year, the industry has cooled from last year’s double- InsideChips’ forecast award for his semiconductor predictions digit pace. The average forecast compiled from all analysts is a for 2002. modest 1.4%. Prior to forming IC Insights, McClean worked at ICE Readers can view the results of last year’s forecasts on Corporation (acquired by Chipworks).
    [Show full text]
  • Corporate Backgrounder
    AGEIA PhysX FAQs for System and Board Partners For use in partner marketing materials KEY QUESTIONS 1: What exactly is the AGEIA PhysX Processor? 2: What is the Gaming Power Triangle? 3: What is physics for gaming? 4: What is a physics processor? 5: What makes the AGEIA PhysX hardware special? 6: Why do I need physics in my game? 7: Isn’t physics just juiced up graphics? 8: Can I get PhysX enabled games today? 9: Why do I need dedicated hardware for physics? 10: How do I get the new AGEIA PhysX processor into my PC? 11: How much does AGEIA PhysX cost? 12: How do games get PhysX-optimized? 13: Do I have to buy new hardware every year just to get new physics features? AGEIA Confidential March 2006 . 1 What exactly is the AGEIA PhysX Processor? The AGEIA PhysX processor is the first dedicated hardware accelerator for PC games that enables dynamic motion and interaction for a totally new gaming experience. Exciting immersive physical gaming feature that you can expect to see in next generation cutting edge titles will include groundbreaking physics features such as: * Explosions that cause dust and collateral debris * Characters with complex, jointed geometries for more life-like motion and interaction * Spectacular new weapons with unpredictable effects * Cloth that drapes and tears the way you expect it to * Lush foliage that sways naturally when brushed against * Dense smoke & fog that billow around objects in motion What is the Gaming Power Triangle The gaming triangle represents the three critical PC processors that enable the best possible gaming experience.
    [Show full text]
  • Datasheet GFX-N3A3-V5LMS1
    NVIDIA GT630 D3 1024 VHDCI to 4 HDMI PCIe® ADD-IN BOARD Datasheet GFX-N3A3-V5LMS1 CONTENTS 1. Feature .................................................................................................................. 3 2. Functional Overview ............................................................................................. 4 2.1. GPU Block diagram .................................................................................... 4 2.2. Key Features .............................................................................................. 4 2.3. Memory ..................................................................................................... 5 2.4. Features and Technologies ....................................................................... 5 2.5. Display Support ......................................................................................... 5 2.6. Video ......................................................................................................... 5 2.7. Bus Support Features ................................................................................ 6 3. PIN Assignment and Description………………………………………………………………………..6 3.1. HDMI Connector Pinout ........................................................................... .6 4. Power Specifications ............................................................................................. 7 5. Thermal Specifications .......................................................................................... 7 6. Output configuration and
    [Show full text]
  • 2014 Annual Review Notice of Annual Meeting Proxy Statement and Form 10-K Nvidia Corporation Nvidia
    NVIDIA CORPORATION 2701 San Tomas Expressway Santa Clara, California 95050 WWW.NVIDIA.COM NVIDIA CORPORATION 2014 ANNUAL REVIEW NOTICE OF ANNUAL MEETING PROXY STATEMENT AND FORM 10-K NVIDIA CORPORATION | 2014 ANNUAL REVIEW ANNUAL 2014 © 2014 NVIDIA Corporation. All rights reserved. NVIDIA, the NVIDIA logo, Compute the Cure, CUDA, GeForce, GTX, GeForce Experience, NVIDIA GameStream, NVIDIA GameWorks, NVIDIA G-SYNC, OptiX, Quadro, Tegra, Tesla, GRID, and SHIELD are trademarks and/or registered trademarks of NVIDIA Corporation in the U.S. and other countries. Other company and product names may be trademarks of the respective companies with which they are associated. © 2014 Respawn Entertainment. All other trademarks are property of their respective owners. 87888_NVDA_Cover.indd 1 3/26/14 3:05 PM CORPORATE INFORMATION BOARD OF DIRECTORS OTHER MEMBERS Tommy Lee Senior Vice President, Jen-Hsun Huang OF THE EXECUTIVE TEAM Systems and Application Engineering Co-Founder, President and Chris A. Malachowsky Chief Executive Offi cer Co-Founder, Senior Vice President Deepu Talla NVIDIA Corporation OVERVIEW and NVIDIA Fellow Vice President and General Manager, Robert K. Burgess Mobile Business Unit Jonah M. Alben Independent Consultant THE VISUAL Senior Vice President, Tony Tamasi Tench Coxe GPU Engineering Senior Vice President, Managing Director, Sutter Hill Ventures Content and Technology Michael J. Byron COMPUTING REVOLUTION James C. Gaither Vice President, Finance, and Bob Worrall Managing Director, Sutter Hill Ventures Chief Accounting Offi cer Senior Vice President and Chief Information Offi cer Dawn Hudson Brian Cabrera ACCELERATES Vice Chairman, The Parthenon Group Senior Vice President and INDEPENDENT ACCOUNTANTS General Counsel Harvey C. Jones PricewaterhouseCoopers LLP NVIDIA is the world leader in visual computing — the art and science of using computers to create and Managing Partner, Square Wave Robert Csongor 488 Almaden Boulevard, Suite 1800 San Jose, California 95110 understand images.
    [Show full text]
  • Designing Heterogeneous Many-Core Processors to Provide High Performance Under Limited Chip Power Budget
    DESIGNING HETEROGENEOUS MANY-CORE PROCESSORS TO PROVIDE HIGH PERFORMANCE UNDER LIMITED CHIP POWER BUDGET A Thesis Presented to The Academic Faculty by Dong Hyuk Woo In Partial Fulfillment of the Requirements for the Degree Doctor of Philosophy in the School of Electrical and Computer Engineering Georgia Institute of Technology December 2010 DESIGNING HETEROGENEOUS MANY-CORE PROCESSORS TO PROVIDE HIGH PERFORMANCE UNDER LIMITED CHIP POWER BUDGET Approved by: Dr. Hsien-Hsin S. Lee, Advisor Dr. Sung Kyu Lim School of Electrical and Computer School of Electrical and Computer Engineering Engineering Georgia Institute of Technology Georgia Institute of Technology Dr. Sudhakar Yalamanchili Dr. Milos Prvulovic School of Electrical and Computer School of Computer Science Engineering Georgia Institute of Technology Georgia Institute of Technology Dr. Marilyn Wolf Date Approved: 23 September 2010 School of Electrical and Computer Engineering Georgia Institute of Technology To my family. iii ACKNOWLEDGEMENTS I would like to take this opportunity to thank all those who directly or indirectly helped me in completing my Ph.D. study. First of all, I would like to thank my advisor, Dr. Hsien-Hsin S. Lee, who contin- uously motivated me, patiently listened to me, and often challenged me with critical feedback. I would also like to thank Dr. Sudhakar Yalamanchili, Dr. Marilyn Wolf, Dr. Sung Kyu Lim, and Dr. Milos Prvulovic for volunteering to serve in my commit- tee and reviewing my thesis. I would also like to thank all the MARS lab members, Dr. Weidong Shi, Dr. Taeweon Suh, Dr. Chinnakrishnan Ballapuram, Dr. Mrinmoy Ghosh, Fayez Mo- hamood, Richard Yoo, Dean Lewis, Eric Fontaine, Ahmad Sharif, Pratik Marolia, Vikas Vasisht, Nak Hee Seong, Sungkap Yeo, Jen-Cheng Huang, Abilash Sekar, Manoj Athreya, Ali Benquassmi, Tzu-Wei Lin, Mohammad Hossain, Andrei Bersatti, and and Jae Woong Sim.
    [Show full text]
  • Accelerator Architectures —A Ten-Year Retrospective
    DEPARTMENT: Expert Opinion Accelerator Architectures —A Ten-Year Retrospective Wen-mei Hwu and This article is a ten-year retrospective of the rise of the Sanjay Patel accelerators since the authors coedited a 2008 IEEE University of Illinois at Urbana-Champaign MICRO special issue on Accelerator Architectures. It identifies the most prominent applications using the accelerators to date: high-performance computing, crypto currencies, and machine learning. For the two most popular types of accelerators, GPUs and FPGAs, the article gives a concise overview of the important trends in their compute throughput, memory bandwidth, and system interconnect. The article also articulates the importance of education for growing the adoption of accelerators. It concludes by identifying emerging types of accelerators and making a few predictions for the coming decade. It has been an exciting ten years since we coedited an IEEE MICRO special issue titled “Accelerator Architectures” in 2008, shortly after NVIDIA launched the CUDA architecture for GPU Computing. In the Guest Editor’s Introduction, we stated: As with any emerging area, the technical delimitations of the field are not well established. The challenges and problems associated with acceleration are still formative. The commercial potential is not generally accepted. There is still the question, “If we build it, who will come?” We define accelerator as a separate architectural substructure (on the same chip, on a different die on the same package, on a different chip on the same board) that is architected using a different set of objectives than the base processor, where these objectives are derived from the needs of a special class of applications.
    [Show full text]
  • Page 1 of 4 NVIDIA CORPORATION NVIDIA Physx DRIVER END USER
    NVIDIA CORPORATION NVIDIA® PhysX™ DRIVER END USER LICENSE AGREEMENT This End User License Agreement (the "Agreement") is a legal agreement between you (either individually or an entity) ("You" or "Your") and NVIDIA Corporation ("NVIDIA") regarding the use of the NVIDIA® PhysX™ Driver and any accompanying documentation (collectively, the "Software"). YOU MUST READ AND AGREE TO THE TERMS OF THIS AGREEMENT BEFORE ANY SOFTWARE CAN BE DOWNLOADED OR INSTALLED OR USED. BY CLICKING ON THE "AGREE" BUTTON OF THIS AGREEMENT, OR INSTALLING SOFTWARE, OR USING SOFTWARE, YOU ARE AGREEING TO BE BOUND BY THE TERMS AND CONDITIONS OF THIS AGREEMENT. IF YOU DO NOT AGREE WITH THE TERMS AND CONDITIONS OF THIS AGREEMENT, THEN YOU SHOULD EXIT THIS PAGE, NOT INSTALL OR USE ANY SOFTWARE, AND DESTROY ALL COPIES OF THE SOFTWARE THAT YOU HAVE DOWNLOADED. BY DOING SO YOU FOREGO ANY IMPLIED OR STATED RIGHTS TO DOWNLOAD OR INSTALL OR USE SOFTWARE. NVIDIA MAY MODIFY THE TERMS OF THIS AGREEMENT FROM TIME TO TIME. ANY USE OF THE PHYSX SDK WILL BE SUBJECT TO SUCH UPDATED TERMS. A CURRENT VERSION OF THIS AGREEMENT IS POSTED ON NVIDIA’S DEVELOPER WEBSITE: www.developer.nvidia.com/object/physx_eula.html This license is only granted to and only may be used by You. NVIDIA grants You a limited, non- exclusive, non-transferable license to use the provided Software for evaluation, testing and production purposes according to the terms set forth below: 1. Use of the Software. a. You may use, display and reproduce the NVIDIA PhysX Driver on Licensed Platforms only. For purposes of this Agreement, “Licensed Platforms” shall include the following: - Any PC or Apple Mac computer with a NVIDIA CUDA-enabled processor executing NVIDIA PhysX; - Any PC or Apple Mac computer running NVIDIA PhysX software executing on the primary central processing unit of the PC only; - Any PC utilizing an AGEIA PhysX processor executing NVIDIA PhysX code; - Microsoft XBOX 360™; - Nintendo® Wii™; and/or - Sony Playstation®3 b.
    [Show full text]