Modeling Instruction Placement on a Spatial Architecture

Total Page:16

File Type:pdf, Size:1020Kb

Modeling Instruction Placement on a Spatial Architecture Modeling Instruction Placement on a Spatial Architecture Martha Mercaldi Steven Swanson Andrew Petersen Andrew Putnam Andrew Schwerin Mark Oskin Susan J. Eggers Computer Science & Engineering University of Washington Seattle, WA USA {mercaldi,swanson,petersen,aputnam,schwerin,oskin,eggers}@cs.washington.edu ABSTRACT Keywords In response to current technology scaling trends, architects dataflow, instruction placement, spatial computing are developing a new style of processor, known as spatial computers. A spatial computer is composed of hundreds or even thousands of simple, replicated processing elements 1. INTRODUCTION (or PEs), frequently organized into a grid. Several current Today’s manufacturing technologies provide an enormous spatial computers, such as TRIPS, RAW, SmartMemories, quantity of computational resources. Computer architects nanoFabrics and WaveScalar, explicitly place a program’s are currently exploring how to convert these resources into instructions onto the grid. improvements in application performance. Despite signifi- Designing instruction placement algorithms is an enor- cant differences in execution models and underlying process mous challenge, as there are an exponential (in the size of technology, five recently proposed architectures - nanoFab- the application) number of different mappings of instruc- rics [18], TRIPS [34], RAW [23], SmartMemories [26], and tions to PEs, and the choice of mapping greatly affects pro- WaveScalar [39] - share the task of mapping large portions gram performance. In this paper we develop an instruction of an application’s binary onto a collection of processing el- placement performance model which can inform instruction ements. Once mapped, the instructions execute “in place”, placement. The model comprises three components, each explicitly sending data between the processing elements. Re- of which captures a different aspect of spatial computing searchers call this form of computation distributed ILP [34, performance: inter-instruction operand latency, data cache 23, 39] or spatial computing [18]. coherence overhead, and contention for processing element Good instruction placement is critical to spatial comput- resources. We evaluate the model on one spatial computer, ing performance. Our research on WaveScalar indicates that WaveScalar, and find that predicted and actual performance a poor placement can decrease performance by as much as a correlate with a coefficient of −0.90. We demonstrate the factor of five. Finding a good placement is hard, because model’s utility by using it to design a new placement al- there are an exponential (in the size of the application) gorithm, which outperforms our previous algorithms. Al- number of possible mappings. How can developers, com- though developed in the context of WaveScalar, the model piler writers, or microarchitects identify the ones that will can serve as a foundation for tuning code, compiling soft- execute quickly? Searching this enormous space requires a ware, and understanding the microarchitectural trade-offs of solid understanding of how instruction placement influences spatial computers in general. performance. In this paper we develop a model of placement performance to study this issue. Categories and Subject Descriptors To develop the model, we focus on a particular spatial computer, WaveScalar. To accurately predict instruction I.6.5 [Computing Methodologies]: Simulation and Mod- placement performance, we construct a unified model that eling—Model Development; B.8.2 [Hardware]: Perfor- considers several factors that contribute to overall perfor- mance and Reliability—Performance Analysis and Design mance. Our model comprises three separate components, Aids each of which captures a different aspect of spatial com- putation: inter-instruction operand latency, data cache co- General Terms herence overhead, and contention for processing element re- Experimentation, Measurement, Performance sources. Our unified model combines these components in proportion to their relative contribution to overall perfor- mance. The model estimates performance using three inputs: (1) Permission to make digital or hard copies of all or part of this work for the placement in question, i.e., a mapping of instructions personal or classroom use is granted without fee provided that copies are in the application to processing elements, (2) a profile of not made or distributed for profit or commercial advantage and that copies application execution behavior, and (3) the spatial com- bear this notice and the full citation on the first page. To copy otherwise, to puter’s microarchitectural configuration and timing param- republish, to post on servers or to redistribute to lists, requires prior specific eters. These inputs are common to all spatial comput- permission and/or a fee. SPAA’06, July 30–August 2, 2006, Cambridge, Massachusetts, USA. ers, which will allow this approach to generalize beyond Copyright2006ACM1-59593-452-9/06/0007...$5.00. WaveScalar. 158 The paper first develops a model of each component of PE placement performance in isolation. Using a variety of appli- Pod Domain cations and potential placements, we evaluate each of these component models, using specially configured versions of D$ the WaveScalar microarchitectural simulator. Each config- uration accurately simulates the hardware resources of the Net- Cluster component in question but idealizes all other resources. We work D$ D$ validate each component model by showing that it correlates S with its component-isolating simulation. L2 L2 L2 B We then combine these component models to pro- duce a single unified model of placement performance on L2 D$ WaveScalar. The unified model predicts the effect of in- struction placement when all microarchitectural resources L2 are accurately simulated. The combined model produces performance predictions that correlate to simulation perfor- mance with a coefficient of −0.90. L2 To evaluate our model’s predictive power on applications that are not part of our workload, we use a standard machine Figure 1: The WaveScalar Processor: The hierar- learning evaluation technique in which we partition our data chical organization of the WaveScalar processor. points into training and test sets. We derive a model from each of the training sets, and evaluate its predictive capabil- tion which computes a value and sends it to the instructions ity on its corresponding test set. Evaluated in this way, our that consume it. Instructions execute after all input operand model’s predicted layout performance correlates to actual values have arrived, according to a principle known as the performance with a coefficient of −0.82. dataflow firing rule [12, 11]. The model indicates that PE resource constraints have WaveScalar supports a memory model which commits the greatest effect on placement performance on WaveScalar, memory accesses in program order. Equipped with archi- followed by inter-instruction operand latency, and finally by tectural building blocks, called waves, which globally order cache coherence overhead. These results are useful in several pieces of the control flow graph, and an architectural mech- ways. For example, the model provides a quickly calculable anism, called wave-ordered memory, which orders memory objective function that an optimizer could minimize to find operations within a wave, WaveScalar enforces the correct, an application mapping that maximizes IPC. One could also global ordering of a thread’s memory operations. This en- use the model to design an instruction placement algorithm ables it to execute applications written in imperative lan- which is based on the factors that are most important to guages, such as C or C++. Other work describes the details performance. In Section 6 we do just this and develop an of this mechanism [39]. improved placement algorithm by combining two existing al- Microarchitecture: Conceptually, each static instruction gorithms that optimize for the two most important compo- in a WaveScalar program executes in a separate processing nents of placement performance, as dictated by the model. element (PE). Building a PE for each static instruction is A third strategy is to use the model to guide microarchi- both impossible and wasteful, so, in practice, WaveScalar tectural optimizations or to make the microarchitecture less dynamically binds multiple instructions to a fixed number placement-sensitive. of PEs, and swaps them in and out on demand. In the following section we provide an overview of the The WaveScalar processor is a grid of simple processing salient features of WaveScalar. In Section 3 we present the elements. Each PE has five pipeline stages and contains methodology used to develop and validate our placement a functional unit, specialized memories to hold operands, performance model. Section 4 explains and validates each and logic to control instruction execution and communica- of the individual components, and Section 5 combines them tion. Each PE also contains buffering and storage for several into a unified model. Section 6 describes an improved in- different static instructions, although only one can execute struction placement algorithm we developed that is based in any given cycle. PEs determine locally when their in- on this model. Section 7 explores related work on perfor- structions can execute, contributing to the scalability of the mance modeling, layout of computation, and spatial com- WaveScalar processor design. puters. Finally in Section 8, we draw our conclusions and To reduce communication costs within
Recommended publications
  • Performance and Energy Efficient Network-On-Chip Architectures
    Linköping Studies in Science and Technology Dissertation No. 1130 Performance and Energy Efficient Network-on-Chip Architectures Sriram R. Vangal Electronic Devices Department of Electrical Engineering Linköping University, SE-581 83 Linköping, Sweden Linköping 2007 ISBN 978-91-85895-91-5 ISSN 0345-7524 ii Performance and Energy Efficient Network-on-Chip Architectures Sriram R. Vangal ISBN 978-91-85895-91-5 Copyright Sriram. R. Vangal, 2007 Linköping Studies in Science and Technology Dissertation No. 1130 ISSN 0345-7524 Electronic Devices Department of Electrical Engineering Linköping University, SE-581 83 Linköping, Sweden Linköping 2007 Author email: [email protected] Cover Image A chip microphotograph of the industry’s first programmable 80-tile teraFLOPS processor, which is implemented in a 65-nm eight-metal CMOS technology. Printed by LiU-Tryck, Linköping University Linköping, Sweden, 2007 Abstract The scaling of MOS transistors into the nanometer regime opens the possibility for creating large Network-on-Chip (NoC) architectures containing hundreds of integrated processing elements with on-chip communication. NoC architectures, with structured on-chip networks are emerging as a scalable and modular solution to global communications within large systems-on-chip. NoCs mitigate the emerging wire-delay problem and addresses the need for substantial interconnect bandwidth by replacing today’s shared buses with packet-switched router networks. With on-chip communication consuming a significant portion of the chip power and area budgets, there is a compelling need for compact, low power routers. While applications dictate the choice of the compute core, the advent of multimedia applications, such as three-dimensional (3D) graphics and signal processing, places stronger demands for self-contained, low-latency floating-point processors with increased throughput.
    [Show full text]
  • Computer Architecture: Dataflow (Part I)
    Computer Architecture: Dataflow (Part I) Prof. Onur Mutlu Carnegie Mellon University A Note on This Lecture n These slides are from 18-742 Fall 2012, Parallel Computer Architecture, Lecture 22: Dataflow I n Video: n http://www.youtube.com/watch? v=D2uue7izU2c&list=PL5PHm2jkkXmh4cDkC3s1VBB7- njlgiG5d&index=19 2 Some Required Dataflow Readings n Dataflow at the ISA level q Dennis and Misunas, “A Preliminary Architecture for a Basic Data Flow Processor,” ISCA 1974. q Arvind and Nikhil, “Executing a Program on the MIT Tagged- Token Dataflow Architecture,” IEEE TC 1990. n Restricted Dataflow q Patt et al., “HPS, a new microarchitecture: rationale and introduction,” MICRO 1985. q Patt et al., “Critical issues regarding HPS, a high performance microarchitecture,” MICRO 1985. 3 Other Related Recommended Readings n Dataflow n Gurd et al., “The Manchester prototype dataflow computer,” CACM 1985. n Lee and Hurson, “Dataflow Architectures and Multithreading,” IEEE Computer 1994. n Restricted Dataflow q Sankaralingam et al., “Exploiting ILP, TLP and DLP with the Polymorphous TRIPS Architecture,” ISCA 2003. q Burger et al., “Scaling to the End of Silicon with EDGE Architectures,” IEEE Computer 2004. 4 Today n Start Dataflow 5 Data Flow Readings: Data Flow (I) n Dennis and Misunas, “A Preliminary Architecture for a Basic Data Flow Processor,” ISCA 1974. n Treleaven et al., “Data-Driven and Demand-Driven Computer Architecture,” ACM Computing Surveys 1982. n Veen, “Dataflow Machine Architecture,” ACM Computing Surveys 1986. n Gurd et al., “The Manchester prototype dataflow computer,” CACM 1985. n Arvind and Nikhil, “Executing a Program on the MIT Tagged-Token Dataflow Architecture,” IEEE TC 1990.
    [Show full text]
  • Configurable Fine-Grain Protection for Multicore Processor Virtualization 1
    Configurable Fine-Grain Protection for Multicore Processor Virtualization David Wentzlaff1, Christopher J. Jackson2, Patrick Griffin3, and Anant Agarwal2 [email protected], [email protected], griffi[email protected], [email protected] 1Princeton University 2Tilera Corp. 3Google Inc. Abstract TLB Access DMA Engine “User” Network I/O Network 2 2 2 2 1 3 1 3 1 3 1 3 Multicore architectures, with their abundant on-chip re- 0 0 0 0 sources, are effectively collections of systems-on-a-chip. The protection system for these architectures must support Key: 0 – User Code, 1 – OS, 2 – Hypervisor, 3 – Hypervisor Debugger multiple concurrently executing operating systems (OSes) Figure 1. With CFP, system software can dy• with different needs, and manage and protect the hard- namically set the privilege level needed to ac• ware’s novel communication mechanisms and hardware cess each fine•grain processor resource. features. Traditional protection systems are insufficient; they protect supervisor from user code, but typically do not protect one system from another, and only support fixed as- ticore systems, a protection system must both temporally signment of resources to protection levels. In this paper, protect and spatially isolate access to resources. Spatial iso- we propose an alternative to traditional protection systems lation is the need to isolate different system software stacks which we call configurable fine-grain protection (CFP). concurrently executing on spatially disparate cores in a mul- CFP enables the dynamic assignment of in-core resources ticore system. Spatial isolation is especially important now to protection levels. We investigate how CFP enables differ- that multicore systems have directly accessible networks ent system software stacks to utilize the same configurable connecting cores to other cores and cores to I/O devices.
    [Show full text]
  • CG-Ooo Energy-Efficient Coarse-Grain Out-Of-Order Execution
    CG-OoO Energy-Efficient Coarse-Grain Out-of-Order Execution Milad Mohammadi⋆, Tor M. Aamodt†, William J. Dally⋆‡ ⋆Stanford University, †University of British Columbia, ‡NVIDIA Research [email protected], [email protected], [email protected] ABSTRACT CG-OoO model to make it even more energy efficient. We introduce the Coarse-Grain Out-of-Order (CG- Despite the significant achievements in improving en- OoO) general purpose processor designed to achieve ergy and performance properties of the OoO proces- close to In-Order processor energy while maintaining sor in the recent years [2], studies show the energy Out-of-Order (OoO) performance. CG-OoO is an and performance attributes of the OoO execution model energy-performance proportional general purpose remain superlinearly proportional [3, 4]. Studies indi- architecture that scales according to the program cate control speculation and dynamic scheduling tech- load1. Block-level code processing is at the heart of nique amount to 88% and 10% of the OoO superior the this architecture; CG-OoO speculates, fetches, performance compared to the In-Order (InO) proces- schedules, and commits code at block-level granu- sor [5]. Scheduling and speculation in OoO is performed larity. It eliminates unnecessary accesses to energy at instruction granularity regardless of the instruction consuming tables, and turns large tables into smaller type even though they are mainly effective during un- and distributed tables that are cheaper to access. predictable dynamic events (e.g. unpredictable cache CG-OoO leverages compiler-level code optimizations misses) [5]. Furthermore, our studies show speculation to deliver efficient static code, and exploits dynamic and dynamic scheduling amount to 67% and 51% of instruction-level parallelism and block-level parallelism.
    [Show full text]
  • Parallel Computer Architecture III
    Parallel Computer Architecture III Stefan Lang Interdisciplinary Center for Scientific Computing (IWR) University of Heidelberg INF 368, Room 532 D-69120 Heidelberg phone: 06221/54-8264 email: [email protected] WS 14/15 Stefan Lang (IWR) Simulation on High-Performance Computers WS 14/15 1 / 51 Parallel Computer Architecture III Parallelism and Granularity Graphic cards I/O Detailed study Hypertransport Protocol Stefan Lang (IWR) Simulation on High-Performance Computers WS 14/15 2 / 51 Parallelism and Granularity level five jobs or Programs coarse grained Subprograms, modules level four or classes middle grained Increase of Increase of Parallelism Communications Procedures, functions level three Requirements or methods Non−recursive loops level two or iterators fine grained Instructions level one or statements Stefan Lang (IWR) Simulation on High-Performance Computers WS 14/15 3 / 51 Graphics Cards GPU = Graphics Processing Unit CUDA = Compute Unified Device Architecture ◮ Toolkit by NVIDIA for direct GPU Programming ◮ Programming of a GPU without graphical API ◮ GPGPU compared to CPUs strongly increased computing performance and storage bandwidth GPUs are cheap and broadly established Stefan Lang (IWR) Simulation on High-Performance Computers WS 14/15 4 / 51 Computing Performance: CPU vs. GPU Stefan Lang (IWR) Simulation on High-Performance Computers WS 14/15 5 / 51 Graphics Cards: Hardware Specification Stefan Lang (IWR) Simulation on High-Performance Computers WS 14/15 6 / 51 Chip Architecture: CPU vs. GPU Stefan Lang (IWR) Simulation on High-Performance Computers WS 14/15 7 / 51 Graphics Cards: Hardware Design Stefan Lang (IWR) Simulation on High-Performance Computers WS 14/15 8 / 51 Graphics Cards: Memory Design 8192 registers (32-bit), in total 32KB per multiprocessor 16KB of fast shared memory per multiprocessor Large global memory (hundreds of MBs, e.g.
    [Show full text]
  • Distributed Microarchitectural Protocols in the TRIPS Prototype Processor
    Øh Appears in the ¿9 Annual International Symposium on Microarchitecture Distributed Microarchitectural Protocols in the TRIPS Prototype Processor Karthikeyan Sankaralingam Ramadass Nagarajan Robert McDonald Ý Rajagopalan DesikanÝ Saurabh Drolia M.S. Govindan Paul Gratz Divya Gulati Heather HansonÝ ChangkyuKim HaimingLiu NityaRanganathan Simha Sethumadhavan Sadia SharifÝ Premkishore Shivakumar Stephen W. Keckler Doug Burger Department of Computer Sciences ÝDepartment of Electrical and Computer Engineering The University of Texas at Austin [email protected] www.cs.utexas.edu/users/cart Abstract are clients on one or more micronets. Higher-level mi- croarchitectural protocols direct global control across the Growing on-chip wire delays will cause many future mi- micronets and tiles in a manner invisible to software. croarchitectures to be distributed, in which hardware re- In this paper, we describe the tile partitioning, micronet sources within a single processor become nodes on one or connectivity, and distributed protocols that provide global more switched micronetworks. Since large processor cores services in the TRIPS processor, including distributed fetch, will require multiple clock cycles to traverse, control must execution, flush, and commit. Prior papers have described be distributed, not centralized. This paper describes the this approach to exploiting parallelism as well as high-level control protocols in the TRIPS processor, a distributed, tiled performanceresults [15, 3], but have not described the inter- microarchitecture that supports dynamic execution. It de- tile connectivity or protocols. Tiled architectures such as tails each of the five types of reused tiles that compose the RAW [23] use static orchestration to manage global oper- processor, the control and data networks that connect them, ations, but in a dynamically scheduled, distributed archi- and the distributed microarchitectural protocols that imple- tecture such as TRIPS, hardware protocols are required to ment instruction fetch, execution, flush, and commit.
    [Show full text]
  • An Evaluation of the TRIPS Computer System
    Appears in the Proceedings of the 14th International Conference on Architecture Support for Programming Languages and Operating Systems An Evaluation of the TRIPS Computer System Mark Gebhart Bertrand A. Maher Katherine E. Coons Jeff Diamond Paul Gratz Mario Marino Nitya Ranganathan Behnam Robatmili Aaron Smith James Burrill Stephen W. Keckler Doug Burger Kathryn S. McKinley Department of Computer Sciences The University of Texas at Austin [email protected] www.cs.utexas.edu/users/cart Abstract issue-width scaling of conventional superscalar architec- The TRIPS system employs a new instruction set architec- tures. Because of these trends, major microprocessor ven- ture (ISA) called Explicit Data Graph Execution (EDGE) dors have abandoned architectures for single-thread perfor- that renegotiates the boundary between hardware and soft- mance and turned to the promise of multiple cores per chip. ware to expose and exploit concurrency. EDGE ISAs use a While many applications can exploit multicore systems, this block-atomic execution model in which blocks are composed approach places substantial burdens on programmers to par- of dataflow instructions. The goal of the TRIPS design is allelize their codes. Despite these trends, Amdahl’s law dic- to mine concurrency for high performance while tolerating tates that single-thread performance will remain key to the emerging technology scaling challenges, such as increas- future success of computer systems [9]. ing wire delays and power consumption. This paper eval- In response to semiconductor scaling trends, we designed uates how well TRIPS meets this goal through a detailed a new architecture and microarchitecture intended to extend ISA and performance analysis.
    [Show full text]
  • A Survey on Coarse-Grained Reconfigurable Architectures from a Performance Perspective
    Received May 31, 2020, accepted July 13, 2020, date of publication July 27, 2020, date of current version August 20, 2020. Digital Object Identifier 10.1109/ACCESS.2020.3012084 A Survey on Coarse-Grained Reconfigurable Architectures From a Performance Perspective ARTUR PODOBAS 1,2, KENTARO SANO1, AND SATOSHI MATSUOKA1,3 1RIKEN Center for Computational Science, Kobe 650-0047, Japan 2Department of Computer Science, KTH Royal Institute of Technology, 114 28 Stockholm, Sweden 3Department of Mathematical and Computing Sciences, Tokyo Institute of Technology, Tokyo 152-8550, Japan Corresponding author: Artur Podobas ([email protected]) This work was supported by the New Energy and Industrial Technology Development Organization (NEDO). ABSTRACT With the end of both Dennard’s scaling and Moore’s law, computer users and researchers are aggressively exploring alternative forms of computing in order to continue the performance scaling that we have come to enjoy. Among the more salient and practical of the post-Moore alternatives are reconfigurable systems, with Coarse-Grained Reconfigurable Architectures (CGRAs) seemingly capable of striking a balance between performance and programmability. In this paper, we survey the landscape of CGRAs. We summarize nearly three decades of literature on the subject, with a particular focus on the premise behind the different CGRAs and how they have evolved. Next, we compile metrics of available CGRAs and analyze their performance properties in order to understand and discover knowledge gaps and opportunities for future CGRA research specialized towards High-Performance Computing (HPC). We find that there are ample opportunities for future research on CGRAs, in particular with respect to size, functionality, support for parallel programming models, and to evaluate more complex applications.
    [Show full text]
  • Designing Heterogeneous Many-Core Processors to Provide High Performance Under Limited Chip Power Budget
    DESIGNING HETEROGENEOUS MANY-CORE PROCESSORS TO PROVIDE HIGH PERFORMANCE UNDER LIMITED CHIP POWER BUDGET A Thesis Presented to The Academic Faculty by Dong Hyuk Woo In Partial Fulfillment of the Requirements for the Degree Doctor of Philosophy in the School of Electrical and Computer Engineering Georgia Institute of Technology December 2010 DESIGNING HETEROGENEOUS MANY-CORE PROCESSORS TO PROVIDE HIGH PERFORMANCE UNDER LIMITED CHIP POWER BUDGET Approved by: Dr. Hsien-Hsin S. Lee, Advisor Dr. Sung Kyu Lim School of Electrical and Computer School of Electrical and Computer Engineering Engineering Georgia Institute of Technology Georgia Institute of Technology Dr. Sudhakar Yalamanchili Dr. Milos Prvulovic School of Electrical and Computer School of Computer Science Engineering Georgia Institute of Technology Georgia Institute of Technology Dr. Marilyn Wolf Date Approved: 23 September 2010 School of Electrical and Computer Engineering Georgia Institute of Technology To my family. iii ACKNOWLEDGEMENTS I would like to take this opportunity to thank all those who directly or indirectly helped me in completing my Ph.D. study. First of all, I would like to thank my advisor, Dr. Hsien-Hsin S. Lee, who contin- uously motivated me, patiently listened to me, and often challenged me with critical feedback. I would also like to thank Dr. Sudhakar Yalamanchili, Dr. Marilyn Wolf, Dr. Sung Kyu Lim, and Dr. Milos Prvulovic for volunteering to serve in my commit- tee and reviewing my thesis. I would also like to thank all the MARS lab members, Dr. Weidong Shi, Dr. Taeweon Suh, Dr. Chinnakrishnan Ballapuram, Dr. Mrinmoy Ghosh, Fayez Mo- hamood, Richard Yoo, Dean Lewis, Eric Fontaine, Ahmad Sharif, Pratik Marolia, Vikas Vasisht, Nak Hee Seong, Sungkap Yeo, Jen-Cheng Huang, Abilash Sekar, Manoj Athreya, Ali Benquassmi, Tzu-Wei Lin, Mohammad Hossain, Andrei Bersatti, and and Jae Woong Sim.
    [Show full text]
  • Compiling for EDGE Architectures
    Appears in the Proceedings of the 4th International Symposium on Code Generation and Optimization (CGO 04). Compiling for EDGE Architectures AaronSmith JimBurrill1 Jon Gibson Bertrand Maher Nick Nethercote BillYoder DougBurger KathrynS.McKinley Department of Computer Sciences 1Department of Computer Science The University of Texas at Austin University of Massachusetts Austin, Texas 78712 Amherst, Massachusetts 01003 Abstract sibilities between programmer, compiler, and hardware to discover and exploit concurrency. Explicit Data Graph Execution (EDGE) architectures of- In previous solutions, CISC processors intentionally fer the possibility of high instruction-level parallelism with placed few ISA-imposed requirements on the compiler to energy efficiency. In EDGE architectures, the compiler expose concurrency. In-order RISC processors required the breaks a program into a sequence of structured blocks that compiler to schedule instructions to minimize pipeline bub- the hardware executes atomically. The instructions within bles for effective pipelining concurrency. With the advent each block communicate directly, instead of communicating of large-window out-of-order microarchitectures, however, through shared registers. The TRIPS EDGE architecture both RISC and CISC processors rely mostly on the hard- imposes restrictions on its blocks to simplify the microar- ware to support superscalar issue. These processors use a chitecture: each TRIPS block has at most 128 instructions, dynamic placement, dynamic issue execution model that re- issues at most 32 loads and/or stores, and executes at most quires the hardware to construct the program dataflow graph 32 register bank reads and 32 writes. To detect block com- on the fly, with little compiler assistance. VLIW processors, pletion, each TRIPS block must produce a constant number conversely, place most of the burden of identifying con- of outputs (stores and register writes) and a branch deci- current instructions on the compiler, which must fill long sion.
    [Show full text]
  • Scatter-Add in Data Parallel Architectures
    Scatter-Add in Data Parallel Architectures Jung Ho Ahn, Mattan Erez and William J. Dally ∗ Computer Systems Laboratory Stanford University, Stanford, CA 94305, USA {gajh,merez,billd}@cva.stanford.edu Abstract histogram Many important applications exhibit large amounts of data parallelism, and modern computer systems are de- signed to take advantage of it. While much of the com- putation in the multimedia and scientific application do- mains is data parallel, certain operations require costly bins serialization that increase the run time. Examples in- clude superposition type updates in scientific computing and histogram computations in media processing. We in- troduce scatter-add, which is the data-parallel form of dataset the well-known scalar fetch-and-op, specifically tuned for SIMD/vector/stream style memory systems. The scatter-add Figure 1: Parallel histogram computation leads to mem- mechanism scatters a set of data values to a set of mem- ory collision, when multiple elements of the dataset up- ory addresses and adds each data value to each refer- date the same histogram bin. enced memory location instead of overwriting it. This novel architecture extension allows us to efficiently sup- tencies. In this paper we will concentrate on the single in- port data-parallel atomic update computations found in struction multiple data (SIMD) class of DPAs, exemplified parallel programming languages such as HPF, and ap- by vector [9, 6], and stream processors [21, 10, 37]. plies both to single-processor and multi-processor SIMD While much of the computation of a typical multimedia data-parallel systems. We detail the micro-architecture ofa or scientific application is indeed data parallel, some sec- scatter-add implementation on a stream architecture, which tions of the code require serialization which significantly requires less than 2% increase in die area yet shows per- limits overall performance.
    [Show full text]
  • Universal Mechanisms for Data-Parallel Architectures
    Universal Mechanisms for Data-Parallel Architectures Karthikeyan Sankaralingam Stephen W. Keckler William R. Mark Doug Burger Computer Architecture and Technology Laboratory Department of Computer Sciences The University of Texas at Austin [email protected] Abstract as high as 20 GFlops (at 450Mhz) on programmable hard- ware [6], which suggests they have at least forty 32-bit Data-parallel programs are both growing in importance floating point units. Software radios for 3G wireless base- and increasing in diversity, resulting in specialized proces- band receivers are being developed for digital signal sors targeted at specific classes of these programs. This pa- processors and require 15 Gops to deliver adequate per- per presents a classification scheme for data-parallel formance [29]. Each arithmetic processor in the Earth program attributes, and proposes micro-architectural Simulator contains forty eight vector pipelines and de- mechanisms to support applications with diverse behav- livers peak performance of up to 8 GFlops. While these ior using a single reconfigurable architecture. We focus domains of data-parallel applications have many com- on the following four broad kinds of data-parallel pro- mon characteristics, they typically show differences in grams — DSP/multimedia, scientific, networking, and the types of memory accesses, computation require- real-time graphics workloads. While all of these pro- ments, and control behavior. grams exhibit high computational intensity, coarse-grain Most data-parallel architectures target a subset of data- regular control behavior, and some regular memory ac- parallel programs, and have poor support for applications cess behavior, they show wide variance in the computation outside of that subset.
    [Show full text]