Lost in Abstraction: Pitfalls of Analyzing Gpus at the Intermediate Language Level

Total Page:16

File Type:pdf, Size:1020Kb

Lost in Abstraction: Pitfalls of Analyzing Gpus at the Intermediate Language Level 2018 IEEE International Symposium on High Performance Computer Architecture Lost in Abstraction: Pitfalls of Analyzing GPUs at the Intermediate Language Level Anthony Gutierrez, Bradford M. Beckmann, Alexandru Dutu, Joseph Gross, John Kalamatianos, Onur Kayiran, Michael LeBeane, Matthew Poremba, Brandon Potter, Sooraj Puthoor, Matthew D. Sinclair, Mark Wyse, Jieming Yin, Xianwei Zhang, Akshay Jain†, Timothy G. Rogers† AMD Research, Advanced Micro Devices, Inc. †School of Electrical and Computer Engineering, Purdue University {anthony.gutierrez, brad.beckmann, alexandru.dutu, joe.gross, john.kalamatianos, onur.kayiran, michael.lebeane, matthew.poremba, brandon.potter, sooraj.puthoor, matthew.sinclair, mark.wyse, jieming.yin, xianwei.zhang}@amd.com {jain156, timrogers}@purdue.edu Abstract—Modern GPU frameworks use a two-phase handful of open-source simulators, such as GPGPU-Sim compilation approach. Kernels written in a high-level [11], gem5 [14], and Multi2Sim [35]. language are initially compiled to an implementation- Simulating a GPU is especially challenging because of agnostic intermediate language (IL), then finalized to the two-phase compilation flow typically used to generate the machine ISA only when the target GPU hardware GPU kernel binaries. To maintain portability between dif- is known. Most GPU microarchitecture simulators ferent generations of GPU hardware, GPU kernels are first available to academics execute IL instructions because compiled to an intermediate language (IL). Prior to launch- there is substantially less functional state associated ing a kernel, low-level software, such as the GPU driver or with the instructions, and in some situations, the ma- runtime, finalizes the IL into the machine instruction set chine ISA’s intellectual property may not be publicly architecture (ISA) representation that targets the GPU disclosed. In this paper, we demonstrate the pitfalls of hardware architecture. evaluating GPUs using this higher-level abstraction, Unlike CPU ISAs, GPU ISAs are often proprietary, and make the case that several important microarchi- change frequently, and require the simulation model to in- tecture interactions are only visible when executing corporate more functional state, thus academic researchers lower-level instructions. often use IL simulation. In particular, NVIDIA’s Parallel Our analysis shows that given identical application Thread Execution (PTX) ISA [33] and the HSA founda- source code and GPU microarchitecture models, execu- tion’s Heterogeneous Systems Architecture Intermediate tion behavior will differ significantly depending on the Language (HSAIL) Virtual ISA [26] are the most popular instruction set abstraction. For example, our analysis ILs. However, AMD has recently disclosed several GPU shows the dynamic instruction count of the machine ISA specifications, including the Graphics Core Next ISA is nearly 2× that of the IL on average, but conten- (GCN) 3 ISA [3] explored in this work. AMD has also re- tion for vector registers is reduced by 3× due to the op- leased their complete GPU software stack under an open- timized resource utilization. In addition, our analysis source license [8], thus it is now feasible for academic re- highlights the deficiencies of using IL to model instruc- searchers to simulate at the lower machine ISA level. tion fetching, control divergence, and value similarity. From an industrial perspective, this paper makes the Finally, we show that simulating IL instructions adds case that using IL execution is not sufficient when evaluat- 33% error as compared to the machine ISA when com- ing many aspects of GPU behavior and implores academics paring absolute runtimes to real hardware. to use machine ISA execution, especially when evaluating microarchitecture features. The primary goals of ILs (e.g., Keywords-ABI; GPU; Intermediate Language; abstracting away hardware details and making kernels Intermediate Representation; ISA; Simulation; portable across different HW architectures) directly con- flict with the goals of cycle-level simulation, such as ac- INTRODUCTION counting for hardware resource contention and taking ad- Research in GPU microarchitecture has increased dra- vantage of unique hardware features. Furthermore, GPUs matically with the advent of GPGPUs. Heterogeneous pro- are co-designed hardware-software systems where vendors gramming features are now appearing in the most popular frequently change the machine ISA to simplify the micro- programming languages, such as C++ [24] and Python architecture and push more complexity to software. [18]. To evaluate research ideas for these massively paral- While several prior works have investigated modeling lel architectures, academic researchers typically rely on cy- challenges for the CPU and memory [12] [15] [20] [23], cle-level simulators. Due to the long development time re- this paper is the first to deeply investigate the unique issues quired to create and maintain these cycle-level simulators, that occur when modeling GPU execution using the IL. To much of the academic research community relies on a facilitate the investigation, this paper also introduces a new simulation infrastructure capable of simulating both 2378-203X/18/$31.00 ©2018 IEEE 608 DOI 10.1109/HPCA.2018.00058 Figure 1: Average of dissimilar and similar statistics. HSAIL and AMD’s GCN3 machine ISA using the same microarchitecture model. Figure 2: Block diagram of the compute unit pipeline. This paper explores the space of statistics for which IL simulation provides a faithful representation of the pro- tomation tool is responsible for identifying the computa- gram under test, and those for which it does not. Figure 1 tion forming the kernel. Modern GPUs support control summarizes several key statistics that differ significantly flow operations allowing programmers to construct com- between HSAIL and GCN3 and a few that do not. Specifi- plex kernels. High-level GPU programming languages de- cally, due to significant code expansion, substantial differ- fine kernels using the single instruction, multiple thread ences in the application binary interface (ABI), and the lack (SIMT) execution model. Threads in a kernel are subdi- of scalar instructions, HSAIL substantially underestimates vided into workgroups, which are further subdivided into the dynamic instruction count and the code footprint. De- wavefronts (WF). All work-items in a WF are executed in spite the significant underestimation in the instruction issue lock step on the SIMD units of a compute unit (CU). Ad- rate, HSAIL substantially overestimates GPU cycles, bank ditionally, AMD’s GCN3 architecture includes a scalar conflicts, value uniqueness within the vector register file unit with corresponding scalar instructions. The use of (VRF), and instruction buffer (IB) flushes. Meanwhile, the these scalar instructions is transparent to the high-level ap- lack of a scalar pipeline does not impact HSAIL’s accuracy plication programmer, as they are generated by the compi- in estimating single instruction, multiple data (SIMD) unit lation toolchain and are inserted into the WF’s instruction utilization and the program’s data footprint. stream, which also includes vector instructions. In summary, we make the following contributions: B. GPU Hardware We add support for the Radeon Open Compute plat- CUs are the scalable execution units that are instanti- form (ROCm) [8] and the state-of-the-art GCN3 ISA ated many times within a GPU. GPUs also typically in- to gem5’s GPU compute model, and we demonstrate clude a multi-level cache hierarchy to improve memory how using the actual runtimes and ABIs impact accu- bandwidth and command processors (CP), which are also racy. called packet processors using HSA terminology [25], to We perform the first study quantifying the effects of manage work assignment. simulating GPUs using an IL, and we identify when Figure 2 provides a high-level block diagram of the CU using an IL is acceptable and when it is not. model in gem5, which is based on AMD’s GCN3 architec- Furthermore, we demonstrate that anticipating the im- ture [2]. Each CU contains four SIMD engines, a scalar pact IL simulation has on estimating runtime is partic- unit, WF slots, local and global memory pipelines, a branch ularly hard to predict and application dependent, thus unit, a scalar register file (SRF), a VRF, a private L1 data architects cannot simply rely on “fudge-factors” to cache, memory coalescing logic, and an on-chip local data make up the difference. share (LDS). Each CU is connected to shared L1 scalar data and instruction caches, which connect to memory through BACKGROUND a shared L2 cache. A. GPU Programming The GCN3 design splits the CU into four separate sets GPUs are programmed using a data parallel, streaming of SIMDs, each of which executes a single instruction computation model. In this model a kernel is executed by a (same PC) in lockstep on sixteen lanes. Thus, a single 64- collection of threads (named work-items in the terminol- work-item WF instruction is executed across 4 cycles. The ogy applied in this paper). A programmer, compiler, or au- GCN3 scalar unit’s primary purpose is to handle control flow and to aid address generation. 609 HSAIL GCN3 HSAIL GCN3 # Return Absolute WI ID # Read AQL Pkt #load kernarg at addr %arg1 # mv kernarg base to v[1:2] workitemabsid $v0, 0; s_load_dword s10, s[4:5], 0x04 ld_kernarg $v[0:1], [%arg1] v_mov v1, s6 v_mov v2, s7 # Wait for number of s_loads = 0 s_waitcnt lgkmcnt(0) #load kernarg into v3 flat_load_dword
Recommended publications
  • Gem5, INTEROPERABILITY, and IMPROVING SIMULATOR METHODOLOGY
    gem5, INTEROPERABILITY, AND IMPROVING SIMULATOR METHODOLOGY Jason Lowe-Power [email protected] @jlowepower Outline What is gem5? Recent gem5 features gem5’s future My vision Big news Features coming soon Created at Michigan by students of Steve Reinhardt, principally Nate Binkert. “A tool for simulating systems” 3 4 Created at Michigan by students of Steve Reinhardt, principally Nate Binkert. “A tool for simulating systems” Created at Wisconsin by students of Mark Hill and David Wood. Detailed memory system 5 6 Today today gem5 2011 paper cited over 3000 times 15-20% of ISCA/MICRO/HPCA papers use gem5 Thriving community project Average of 70-ish commits per month About 100 unique contributors over last 2 years 7 Today today Over 400 “models” Over 4000 parameters! 3 timing-based CPU models (simple, in order, out of order) 8 ISAs (ARM, RISC-V, x86, Alpha, Power, SPARC, MIPS, GCN3) 12 memory models (DDR3, DDR4, HBM, HMC, etc.) 42 devices (PCI, Arm platform, x86 platform, storage) AMD GPGPU 12 cache coherence protocols Network on chip (Garnet) 8 Today Programmatic configuration 9 Today Programmatic configuration 10 11 Today today Support for complex ML stacks See https://github.com/KyleRoarty/gem5_docker/ Support for ARM SVE instructions Integration with other simulators DRAMSim2 SST DSENT SystemC McPAT GPGPU-Sim 12 Simulator integration Choose a driver Choose an interface 13 gem5 + SST Multiple implementations SST used as driver for gem5 But two event queues Interface is memory requests Separate setup for each simulator 14 gem5 + SystemC
    [Show full text]
  • SST + Gem5 = a Scalable Simulation Infrastructure for High Performance Computing
    SST + gem5 = A Scalable Simulation Infrastructure for High Performance Computing Mingyu Hsieh Jie Meng Michael Levenhagen Sandia National Labs P.O.Box Boston University Sandia National Labs 5800 ECE Department P.O.Box 5800 Albuquerque, NM Boston, MA Albuquerque, NM [email protected] [email protected] [email protected] Kevin Pedretti Ayse Coskun Arun Rodrigues Sandia National Labs Boston University Sandia National Labs P.O.Box 5800 ECE Department P.O.Box 5800 Albuquerque, NM Boston, MA Albuquerque, NM [email protected] [email protected] [email protected] ABSTRACT 1. INTRODUCTION High Performance Computing (HPC) faces new challenges in As HPC continues to push the bounds of computation, it scalability, performance, reliability, and power consumption. faces new challenges in scalability, performance, reliability, Solving these challenges will require radically new hardware and power consumption. To address these challenges, we and software approaches. It is impractical to explore this need to design new architectures, operating systems, pro- vast design space without detailed system-level simulations. gramming models, and software. It is impractical to ex- However, most of the existing simulators are either not suffi- plore this vast design space without detailed system-level ciently detailed, not scalable, or cannot evaluate key system simulations. However, most simulators are either not suffi- characteristics such as energy consumption or reliability. ciently detailed (i.e., glossing over important details in pro- To address this problem, we integrate the highly detailed cessor, memory, or network implementations), not scalable gem5 performance simulator into the parallel Structural Sim- (i.e., simulating highly parallel architectures with serial sim- ulation Toolkit (SST).
    [Show full text]
  • THE AMD Gem5 APU SIMULATOR: MODELING GPUS USING the MACHINE ISA
    THE AMD gem5 APU SIMULATOR: MODELING GPUS USING THE MACHINE ISA TONY GUTIERREZ, SOORAJ PUTHOOR, TUAN TA*, MATT SINCLAIR, AND BRAD BECKMANN AMD RESEARCH, *CORNELL JUNE 2, 2018 OBJECTIVES AND SCOPE Objectives ‒ Introduce the Radeon Open Compute Platform (ROCm) ‒ AMD’s Graphics Core Next (GCN) architecture and GCN3 ISA ‒ Describe the gem5-based APU simulator Modeling Scope APU ‒ Emphasis on the GPU side of the simulator systems ‒ APU (CPU+GPU) model, not discrete GPU ‒ Covers GPU arch, GCN3 ISA, and HW-SW interfaces Why are we releasing our code? ‒ Encourage AMD-relevant research ‒ Modeling ISA and real system stack is important [1] ‒ Enhance academic collaborations ‒ Enable intern candidates to get experience before arriving ‒ Enable interns to take their experience back to school Acknowledgement ‒ AMD Research’s gem5 team [1] Gutierrez et al. Lost in Abstraction: Pitfalls of Analyzing GPUs at the Intermediate Language Level. HPCA, 2018. 2 | THE AMD gem5 APU SIMULATOR | JUNE 2, 2018 | ISCA 2018 TUTORIAL | QUICK SURVEY Who is in our audience? ‒ Graduate students ‒ Faculty members ‒ Working for government research labs ‒ Working for industry Have you written an GPU program? ‒ CUDA, OpenCLTM, HIP, HC, C++ AMP, other languages Have you used these simulators? ‒ GPGPU-Sim ‒ Multi2Sim ‒ gem5 ‒ Our HSAIL-based APU model Are you familiar with our HPCA 2018 paper? ‒ Lost in Abstraction: Pitfalls of Analyzing GPUs at the Intermediate Language Level 3 | THE AMD gem5 APU SIMULATOR | JUNE 2, 2018 | ISCA 2018 TUTORIAL | OUTLINE Topic Presenter Time Background Tony 8:00 – 8:15 ROCm, GCN3 ISA, and GPU Arch Tony 8:15 – 9:15 HSA Implementation in gem5 Sooraj 9:15 – 10:00 Break 10:00 – 10:30 Ruby and GPU Protocol Tester Tuan 10:30 – 11:15 Demo and Workloads Matt 11:15 – 11:50 Summary and Questions All 11:50 – 12:00 4 | THE AMD gem5 APU SIMULATOR | JUNE 2, 2018 | ISCA 2018 TUTORIAL | BACKGROUND Overview of gem5 ‒ Source tree GPU terminology and system overview HSA standard and building blocks ‒ Coherent shared virtual memory ‒ User-level queues ‒ Signals ‒ etc.
    [Show full text]
  • Enhancing the RISC-V Instruction Set Architecture
    Enhancing the RISC-V Instruction Set Architecture Plamen Petrov MInf Project (Part 1) Report Master of Informatics School of Informatics University of Edinburgh 2019 3 Abstract Since its emergence, RISC-V has offered a license-free alternative to proprietary in- struction sets designed by ARM and Intel by providing a stable base to researchers and small companies. This makes it possible for small companies and researchers to design new cores without incurring exorbitant costs. In our two-part project we explore RISC- V’s performance compared to the commercial ARM and X86 architectures. Most of the first part focuses on static code size, where we find that RISC-V performs very rea- sonably, but is behind Thumb2. For this reason, we propose improvements to RISC-V which constitute of new prologue and epilogue instructions that almost fully bridge this gap. We give ideas for other possible enhancements and plan to extend the project to cover dynamic measures in the second part. Overall, we conclude that RISC-V per- forms very well given the resources invested in it and open instruction sets, in general, deserve to be developed further for the better of the computer architecture field as a whole. 4 Acknowledgements I would like to express my gratitude to my supervisor Professor Nigel Topham for supporting me with the project throughout the year. Table of Contents 1 Introduction 7 1.1 Overview . 7 1.2 Aims . 9 1.3 Contributions . 10 1.4 Outline . 10 2 Background 11 2.1 Abbreviations . 11 2.2 Binary File . 11 2.3 SPEC2006 .
    [Show full text]
  • Latest Status of These Tests on Gem5
    gem5art Release 0.2.1 Jul 01, 2021 Contents 1 Introduction 3 1.1 Installing gem5art............................................4 2 Artifacts 5 2.1 gem5art artifacts.............................................5 2.2 ArtifactDB................................................7 2.3 Artifacts API Documentation......................................9 3 Run 15 3.1 Introduction............................................... 15 3.2 SE and FS mode runs.......................................... 15 3.3 Running an experiment.......................................... 16 3.4 Run Already in the Database....................................... 17 3.5 Searching the Database to find Runs................................... 17 3.6 Searching the Database to find Runs with Specific Names....................... 18 3.7 Runs API Documentation........................................ 18 4 Tasks 21 4.1 Use of Python Multiprocessing..................................... 21 4.2 Use of Celery............................................... 21 4.3 Tasks API Documentation........................................ 22 5 Disk Images 23 5.1 Introduction............................................... 23 5.2 Building a Simple Disk Image with Packer............................... 23 6 Frequently Asked Questions 27 7 Tutorial: Run Full System Linux Boot Tests 29 7.1 Introduction............................................... 29 7.2 Setting up the environment........................................ 30 7.3 Building gem5.............................................. 30 7.4 Creating a disk
    [Show full text]
  • Sources of Error in Full-System Simulation
    Sources of Error in Full-System Simulation Anthony Gutierrez Joseph Pusdesris Chander Sudanthi Christopher D. Emmons Ronald G. Dreslinski Trevor Mudge Mitchell Hayenga Nigel Paver Advanced Computer Architecture Laboratory ARM Research - Austin, TX University of Michigan - Ann Arbor, MI fchander.sudanthi, chris.emmons, mitch.hayenga, [email protected] fatgutier, joemp, rdreslin, [email protected] Abstract—In this work we investigate the sources of error corrected. Specification and abstraction errors, however, may in gem5—a state-of-the-art computer simulator—by validating remain in the simulator for several reasons—certain features it against a real hardware platform: the ARM Versatile Express of a system may not be publicly available, or the abstractions TC2 development board. We design a custom gem5 configuration may simplify the simulator code and improve simulation and make several changes to the simulator itself in order to speed. These abstractions may not impact the accuracy of the more closely match the Versatile Express TC2 board. With simulator significantly for its intended usage scenario, however the modifications we make to the simulator, we are able to achieve a mean percentage runtime error of 5% and a mean they may cause experimental error that is unacceptable when absolute percentage runtime error of 13% for the SPEC CPU2006 the simulator is modified to represent a system it was not benchmarks. For the PARSEC benchmarks we achieve a mean originally intended to model. Therefore, it is necessary to percentage runtime error of -11% and -12% for single and dual- validate the simulator to ensure that these errors do not lead core runs respectively, and a mean absolute percentage runtime to unreliable results.
    [Show full text]
  • Empirical CPU Power Modelling and Estimation in the Gem5 Simulator
    Empirical CPU Power Modelling and Estimation in the gem5 Simulator Basireddy Karunakar Reddy∗, Matthew J. Walker∗, Domenico Balsamo∗, Stephan Diestelhorsty, Bashir M. Al-Hashimi∗ and Geoff V. Merrett∗ ∗University of Southampton yARM Ltd. Southampton, UK Cambridge, UK fkrb1g15,mw9g09,db2a12,bmah,[email protected] [email protected] Abstract—Power modelling is important for modern CPUs The bottom-up power models take a CPU specification (e.g. to inform power management approaches and allow design cache size, number of pipelines) and attempt to estimate the space exploration. Power simulators, combined with a full-system power by first estimating the CPU complexity, number of architectural simulator such as gem5, enable power-performance trade-offs to be investigated early in the design of a system components and area and then the amount of switching activity with different configurations (e.g number of cores, cache size, required to perform different tasks. Wattch is a framework for etc.). However, the accuracy of existing power simulators, such architectural-level power simulation. CACTI is a modelling as McPAT, is known to be low due to the abstraction and tool for estimating cache and memory access time, cycle specification errors, and this can lead to incorrect research time, area, leakage, and dynamic power using device models conclusions. In this paper, we present an accurate power model, built from measured data, integrated into gem5 for estimating based on the industry-standard ITRS roadmap. McPAT is an the power consumption of a simulated quad-core ARM Cortex- integrated power, area, and timing modelling framework. In A15. A power modelling methodology based on Performance particular, McPAT has gained popularity due to its ease-of-use Monitoring Counters (PMCs) is used to build and evaluate the and readiness, as it needs a single configuration file with activ- integrated model in gem5.
    [Show full text]
  • Appendices a Sample Gem5 Python Configuration Code
    Table of Contents I Table of Contents List of Figures III List of Acronyms IV 1 Introduction and background 1 1.1 Real-time systems . ............... 1 1.2 Verification of the real-time systems . 1 1.3 The worst case execution time of a software program . 3 1.3.1 Static analysis . ............. 3 1.3.2 Measurement . ............. 4 1.4 Hardware model in analysis tools . 4 1.5 Moving toward heterogeneous systems . 6 2 Thesis goals and contributions 7 2.1 Goals of the thesis . ................ 7 2.2 Structure of the thesis . .............. 7 3 Related work 8 3.1 Hardware models . ............... 8 3.2 Simulators . ................... 8 4 Gem5 input hardware model 10 5 New hardware description model 13 5.1 Model specification . 13 5.1.1 Processor . 13 5.1.2 Cache . 14 5.1.3 Memory . 15 5.1.4 Link . ................... 16 5.1.5 Group . 16 5.2 Model JSON format . 16 5.3 Model example . 18 6 Application overview 21 6.1 Application functionalities . 21 6.2 How application works . 22 © Uppsala University Duc Anh Nguyen Table of Contents II 6.3 Gem5 simulation and measuring WCET . 25 6.3.1 Gem5 Python configurationfile.................... 25 6.3.2 Translating model to arguments and simulating . 26 6.3.3 Extracting output statistics . 28 7 Conclusion and future work 30 7.1 Future work . 30 7.1.1 The hardware desciption model . 30 7.1.2 The application . 30 References 32 Appendices 35 A Sample Gem5 Python configuration code 35 © Uppsala University Duc Anh Nguyen List of Figures III List of Figures Figure 1.1: Overview of an analysis tool .
    [Show full text]
  • Validation of the Gem5 Simulator for X86 Architectures
    Validation of the gem5 Simulator for x86 Architectures Ayaz Akram1 Lina Sawalha2 University of California, Western Michigan University, Davis, CA Kalamazoo, MI [email protected] [email protected] Abstract—gem5 has been extensively used in computer archi- heterogeneous systems composed of many cores and complex tecture simulations and in the evaluation of new architectures configurations. gem5 has been used by ARM research to for HPC (high performance computing) systems. Previous work perform HPC platform simulation and by AMD for their has validated gem5 against ARM platforms. However, gem5 still shows high inaccuracy when modeling x86 based processors. Exascale project [4]. In this work, we focus on the simulation of a single node gem5 [3] supports several ISAs (instruction set architec- high performance system and study the sources of inaccuracies tures) like x86, ARM, Alpha, MIPS, RISC-V and SPARC. It of gem5. Then we validate gem5 simulator against an Intel supports simple and quick functional simulation and detailed processor, Core-i7 (Haswell microarchitecture). We configured simulation modes. The detailed simulation can be either with gem5 as close as possible to match Core-i7 Haswell microar- chitecture configurations and made changes to the simulator to system call emulation or with full-system support, in which add some features, modified existing code, and tuned built-in gem5 is capable of booting a Linux operating system. gem5 configurations. As a result, we validated the simulator by fixing supports two different pipeline implementations, in-order (IO) many sources of errors to match real hardware results with and out-of-order (OoO). gem5 has been previously validated less than 6% mean error rate for different control, memory, for ARM based targets [5], [6].
    [Show full text]
  • The Gem5 Simulator: Version 20.0+∗ a New Era for the Open-Source Computer Architecture Simulator
    The gem5 Simulator: Version 20.0+∗ A new era for the open-source computer architecture simulator Jason Lowe-Power, Abdul Mutaal Ahmad, Ayaz Akram, Mohammad Alian, Rico Amslinger, Matteo Andreozzi, Adrià Armejach, Nils Asmussen, Brad Beckmann, Srikant Bharadwaj, Gabe Black, Gedare Bloom, Bobby R. Bruce, Daniel Rodrigues Carvalho, Jeronimo Castrillon, Lizhong Chen, Nicolas Derumigny, Stephan Diestelhorst, Wendy Elsasser, Carlos Escuin, Marjan Fariborz, Amin Farmahini-Farahani, Pouya Fotouhi, Ryan Gambord, Jayneel Gandhi, Dibakar Gope, Thomas Grass, Anthony Gutierrez, Bagus Hanindhito, Andreas Hansson, Swapnil Haria, Austin Harris, Timothy Hayes, Adrian Herrera, Matthew Horsnell, Syed Ali Raza Jafri, Radhika Jagtap, Hanhwi Jang, Reiley Jeyapaul, Timothy M. Jones, Matthias Jung, Subash Kannoth, Hamidreza Khaleghzadeh, Yuetsu Kodama, Tushar Krishna, Tommaso Marinelli, Christian Menard, Andrea Mondelli, Miquel Moreto, Tiago Mück, Omar Naji, Krishnendra Nathella, Hoa Nguyen, Nikos Nikoleris, Lena E. Olson, Marc Orr, Binh Pham, Pablo Prieto, Trivikram Reddy, Alec Roelke, Mahyar Samani, Andreas Sandberg, Javier Setoain, Boris Shingarov, Matthew D. Sinclair, Tuan Ta, Rahul Thakur, Giacomo Travaglini, Michael Upton, Nilay Vaish, Ilias Vougioukas, William Wang, Zhengrong Wang, Norbert Wehn, Christian Weis, David A. Wood, Hongil Yoon, Éder F. Zulian † ABSTRACT and every year many papers published in the top computer archi- The open-source and community-supported gem5 simulator is one tecture venues use gem5 as their main evaluation infrastructure. of the most popular tools for computer architecture research. This Additionally, gem5 is one of the tools used to design the Fugaku simulation infrastructure allows researchers to model modern com- supercomputer, one of the first exascale systems [45]. puter hardware at the cycle level, and it has enough fidelity to boot The gem5 simulator [18] is an open source community-supported unmodified Linux-based operating systems and run full applica- computer architecture simulator system.
    [Show full text]
  • Evaluating Gem5 and QEMU Virtual Platforms for ARM Multicore Architectures
    DEGREE PROJECT IN COMPUTER SCIENCE AND ENGINEERING, SECOND CYCLE, 30 CREDITS STOCKHOLM, SWEDEN 2016 Evaluating Gem5 and QEMU Virtual Platforms for ARM Multicore Architectures JOSÉ LUIS BISMARCK FUENTES MORALES KTH ROYAL INSTITUTE OF TECHNOLOGY SCHOOL OF INFORMATION AND COMMUNICATION TECHNOLOGY Evaluating Gem5 and QEMU Virtual Platforms for ARM Multicore Architectures José Luis Bismarck Fuentes Morales 2016-09-25 Master’s Thesis Examiner Ingo Sander Academic adviser Johnny Öberg KTH Royal Institute of Technology School of Information and Communication Technology (ICT) Department of Communication Systems SE-100 44 Stockholm, Sweden Abstract | iii Abstract Accurate virtual platforms allow for crucial, early, and inexpensive assessments about the viability and hardware constraints of software/hardware applications. The growth of multicore architectures in both number of cores and relevance in the industry, in turn, demands the emergence of faster and more efficient virtual platforms to make the benefits of single core simulation and emulation available to their multicore successors whilst maintaining accuracy, development costs, time, and efficiency at acceptable levels. The goal of this thesis is to find optimal virtual platforms to perform hardware design space exploration for multi-core architectures running filtering functions, particularly, a discrete signal filtering Matlab algorithm used for oil surveying applications running on an ARM Cortex-A53 quadcore CPU. In addition to the filtering algorithm, the PARSEC benchmark suite was also used to test platform compliance under workloads with diverse characteristics. Upon reviewing multiple virtual platforms, the gem5 simulator and the QEMU emulator were chosen to be tested due to their ubiquitousness, prominence and flexibility. A Raspberry Pi Model B was used as reference to measure how closely these tools can model a commonly used embedded platform.
    [Show full text]
  • Accelerates In-Memory Databases with Near Data Computing Kevin Hsieh Amirali Boroumand [email protected] [email protected]
    Accelerates In-Memory Databases with Near Data Computing Kevin Hsieh Amirali Boroumand [email protected] [email protected] 1. Problem and Motivation functions to parallelize them and hide the memory latency. Doing these operations on the memory side can minimize Modern databases, such as H-Store[14], VoltDB[23], and the data movements over wires, which costs 23X energy of Masstree[20], have been moving from disk to main memory double-precision fused-multiply add operation in 10nm[15]. because cheaper commodity DRAMs are orders of magnitude Processing bulk data on the memory side can also significantly faster. Furthermore, DRAM-based key-value storage servers, reduce cache thrashing on the processor side. such as memcached[9] and Redis[22], have been known as critical parts of major Internet services. 3. Related Works All the paradigm shifts from disks to DRAMs have made the bandwidth and latency of DRAMs, or memory wall[26], Several prior studies analyzed the bottlenecks of databases and become the emerging bottleneck of data centers. Conventional proposed specialized accelerators to speed them up. The most databases are known to spend at least 50% of their execution common benchmarks are TPC-C for online transaction pro- time on memory stalls[3]. This problem is getting worse for cessing (OLTP) and TPC-H for online analytical processing in-memory databases. For example, the profiling of Masstree (OLAP)[2]. Most OLAP related works used MonetDB[12], showed that the get operations spend 66%-75% of time on an in-memory, column-store database. On the other hand, the DRAM stalls[20].
    [Show full text]