Software Cache Coherence for Large Scale Multiprocessors HPCA

Total Page:16

File Type:pdf, Size:1020Kb

Software Cache Coherence for Large Scale Multiprocessors HPCA Software Cache Coherence for Large Scale Multiprocessors Leonidas I. Kontothanassis and Michael L. Scott Department of Computer Science, University of Rochester, Rochester, NY 14627-0226 ¡ kthanasi,scott ¢ @cs.rochester.edu Abstract Unfortunately, the current state of the art in software Shared memory is an appealing abstraction for paral- coherence for message-passing machines provides perfor- lel programming. It must be implemented with caches in mance nowhere close to that of hardware cache coherence. order to perform well, however, and caches require a coher- Our contributionis to demonstrate that most of the bene®ts ence mechanism to ensure that processors reference current of hardware cache coherence can be obtained on large ma- data. Hardware coherence mechanisms for large-scale ma- chines simplyby providinga globalphysical address space, chines are complex and costly, but existing software mech- with per-processor caches but without hardware cache anisms for message-passing machines have not provided coherence. Machines in this non-cache-coherent, non- a performance-competitive solution. We claim that an in- uniform memory access (NCC-NUMA) class include the termediate hardware optionÐmemory-mapped network in- Cray Research T3D and the Princeton Shrimp [6]. In com- terfaces that support a global physical address spaceÐcan parison to hardware-coherent machines, NCC-NUMAs can providemost oftheperformance bene®ts ofhardware cache more easily be built from commodity parts, and can follow coherence. We present a software coherence protocol that improvements in microprocessors and other hardware tech- runs on this class of machines and greatly narrows the per- nologies closely. formance gap between hardware and software coherence. We present a software coherence protocol for NCC- We compare the performance of the protocol to that of NUMA machines that scales well to large numbers of pro- existing software and hardware alternatives and evaluate cessors. To achieve the best possible performance, we the tradeoffs among various cache-write policies. We also exploit the global address space in three speci®c ways. observe that simple program changes can greatly improve First, we maintain directory information for the coherence performance. For the programs in our test suite and with protocol in uncached shared locations, and access it with the changes in place, software coherence is often faster and ordinary loads and stores, avoiding the need to interrupt never more than 55% slower than hardware coherence. remote processors in almost all circumstances. Second, while using virtual memory to maintain coherence at the 1 Introduction granularity of pages, we never copy pages. Instead, we Large scale multiprocessors can provide the computa- map them remotely and allow the hardware to fetch cache tional power needed for some of the larger problems of lines on demand. Third, while allowing multiple writers science and engineering today. Shared memory provides for concurrency, we avoid the need to keep old copies and an appealing programming model for such machines. To compute diffs by using ordinary hardware write-throughor perform well, however, shared memory requires the use write-back to the unique main-memory copy of each page. of caches, which in turn require a coherence mechanism to Full-scale hardware cache coherence may suffer less ensure that copies of data are up-to-date. Coherence iseasy from false sharing, due to smaller coherence blocks, and to achieve on small, bus-based machines, where every pro- can execute protocol operations in parallel with ªrealº com- cessor can see the memory traf®c of the others [4, 11], but putation. Current trends, however, are reducing the impor- is substantially harder to achieve on large-scale multipro- tance of each of these advantages: relaxed consistency, bet- cessors [1, 15, 19]. Itincreases boththe costof the machine ter compilers, and improvements in programming method- and the time and intellectual effort required to bring it to ology can also reduce false sharing, and trends in both market. Given the speed of advances in microprocessor programming methodology and hardware technology are technology, long development times generally lead to ma- reducing the fraction of total computation time devoted to chines with out-of-date processors. There is thus a strong protocol execution. At the same time, software coherence motivation to ®nd coherence mechanisms that will produce enjoys the advantage of being able to employ techniques acceptable performance with little or no special hardware.1 such as multiple writers, delayed write notices, and delayed invalidations, which may be too complicated to implement £ This work was supported in part by NSF Institutional Infrastructure reliably in hardware at acceptable cost. Software coherence grant no. CDA-8822724 and ONR research grant no. N00014-92-J-1801 also offers the possibility of adapting protocols to individ- (in conjunction with the DARPA Research in Information Science and ual programsor dataregionswitha level of ¯exibilitythatis TechnologyÐHighPerformance Computing, Software Science and Tech- dif®culttoduplicateinhardware. Weexploitthisadvantage nology program, ARPA Order no. 8930). in part in our work by employing a relatively complicated 1We are speaking here of behavior-driven coherenceÐmechanisms eight-state protocol, by using uncached remote references that move and replicate data at run time in response to observed patterns for application-level data structures that are accessed at a of program behaviorÐas opposed to compiler-based techniques [9]. very ®ne grain, and by introducing user level annotations that can impact the behavior of the coherence protocol. the home node. In addition, each processor holds a local The rest of the paper is organized as follows. Section 2 weak list that indicates which of the pages to which it has describes our software coherence protocol and provides in- mappings are weak. When a processor takes a page fault tuition for our algorithmic and architectural choices. Sec- it locks the coherent map entry representing the page on tion 3 describes our experimental methodology and work- which the fault was taken. It then changes the coherent load. We present performance results in section 4 and com- map entry to re¯ect the new state of the page. If necessary pare our work to other approaches in section 5. We compare (i.e. if the page has made the transition from shared or our protocol to a variety of existing alternatives, including dirty to weak) the processor updates the weak lists of all sequentially-consistent hardware, release-consistent hard- processorsthat havemappingsfor that page. It then unlocks ware, straightforwardsequentially-consistentsoftware, and the entry in the coherent map. The process of updating a a coherence scheme for small-scale NCC-NUMAs due to processor's weak list is referred to as posting a write notice. Petersen and Li [21]. We also report on the impact of sev- Use of a distributed coherent map and per-processor eral architectural alternatives on the effectiveness of soft- weak lists enhances scalability by minimizing memory con- ware coherence. These alternatives include the choice of tention and by avoiding the need for processors at acquire write policy (write-through,write-back, write-throughwith points to scan weak list entries in which they have no inter- a write-merge buffer) and the availability of a remote refer- est (something that would happen with a centralized weak ence facility, which allows a processor to choose to access list [20]. However it may make the transition to the weak data directly in a remote location, by disabling caching. state very expensive, since a potentially large number of Finally, to obtain the full bene®t of software coherence, we remote memory operations may have to be performed (se- observe that minor program changes can be crucial. In par- rially) in order to notify all sharing processors. Ideally, ticular, we identify the need to employ reader-writer locks, we wouldlike to maintain the low acquire overhead of per- avoid certain interactions between program synchroniza- processor weak listswhilerequiringonlya constantamount tion and the coherence protocol, and align data structures of work per shared page on a transition to the weak state. with page boundaries whenever possible. We summarize In order to approach this goal we take advantage of the our ®ndings and conclude in section 6. fact that page behavior tends to be relatively constant over the execution of a program, or at least a large portion of 2 The Software Coherence Protocol it. Pages that are weak at one acquire point are likely to In this section we present a scalable protocol for soft- be weak at another. We therefore introduce an additional ware cache coherence. As in most software coherence pair of states, called safe and unsafe. These new states, systems, we use virtual memory protection bits to enforce which are orthogonal to the others (for a total of 8 distinct consistency at the granularity of pages. As in Munin [26], states), re¯ect the past behavior of the page. A page that Treadmarks [14], and the work of Petersen and Li [20, 21], has made the transition to weak several times and is about we allow more than one processor to write a page concur- to be marked weak again is also marked as unsafe. Future rently, and we use a variant of release consistency to limit transitionstothe weak state willno longerrequirethe send- coherence operations to synchronization points. (Between ing of write notices. Instead the processor that causes the these points, processors can continue to use stale data in transition to the weak state changes only the entry in the their caches.) As in the work of Petersen and Li, we exploit coherent map, and then continues. The acquire part of the the global physical address space to move data at the gran- protocol now requires that the acquiring processor check ularity of cache lines: instead of copying pages we map the coherent map entry for all its unsafe pages, and inval- them remotely, and allow the hardware to fetch cache lines idate the ones that are also marked as weak.
Recommended publications
  • A Type Inference on Executables
    A Type Inference on Executables Juan Caballero, IMDEA Software Institute Zhiqiang Lin, University of Texas at Dallas In many applications source code and debugging symbols of a target program are not available, and what we can only access is the program executable. A fundamental challenge with executables is that during compilation critical information such as variables and types is lost. Given that typed variables provide fundamental semantics of a program, for the last 16 years a large amount of research has been carried out on binary code type inference, a challenging task that aims to infer typed variables from executables (also referred to as binary code). In this article we systematize the area of binary code type inference according to its most important dimensions: the applications that motivate its importance, the approaches used, the types that those approaches infer, the implementation of those approaches, and how the inference results are evaluated. We also discuss limitations, point to underdeveloped problems and open challenges, and propose further applications. Categories and Subject Descriptors: D.3.3 [Language Constructs and Features]: Data types and struc- tures; D.4.6 [Operating Systems]: Security and Protection General Terms: Languages, Security Additional Key Words and Phrases: type inference, program executables, binary code analysis ACM Reference Format: Juan Caballero and Zhiqiang Lin, 2015. Type Inference on Executables. ACM Comput. Surv. V, N, Article A (January YYYY), 35 pages. DOI:http://dx.doi.org/10.1145/0000000.0000000 1. INTRODUCTION Being the final deliverable of software, executables (or binary code, as we use both terms interchangeably) are everywhere. They contain the final code that runs on a system and truly represent the program behavior.
    [Show full text]
  • Targeting Embedded Powerpc
    Freescale Semiconductor, Inc. EPPC.book Page 1 Monday, March 28, 2005 9:22 AM CodeWarrior™ Development Studio PowerPC™ ISA Communications Processors Edition Targeting Manual Revised: 28 March 2005 For More Information: www.freescale.com Freescale Semiconductor, Inc. EPPC.book Page 2 Monday, March 28, 2005 9:22 AM Metrowerks, the Metrowerks logo, and CodeWarrior are trademarks or registered trademarks of Metrowerks Corpora- tion in the United States and/or other countries. All other trade names and trademarks are the property of their respective owners. Copyright © 2005 by Metrowerks, a Freescale Semiconductor company. All rights reserved. No portion of this document may be reproduced or transmitted in any form or by any means, electronic or me- chanical, without prior written permission from Metrowerks. Use of this document and related materials are governed by the license agreement that accompanied the product to which this manual pertains. This document may be printed for non-commercial personal use only in accordance with the aforementioned license agreement. If you do not have a copy of the license agreement, contact your Metrowerks representative or call 1-800-377- 5416 (if outside the U.S., call +1-512-996-5300). Metrowerks reserves the right to make changes to any product described or referred to in this document without further notice. Metrowerks makes no warranty, representation or guarantee regarding the merchantability or fitness of its prod- ucts for any particular purpose, nor does Metrowerks assume any liability arising
    [Show full text]
  • Interprocedural Analysis of Low-Level Code
    TECHNISCHE UNIVERSITAT¨ MUNCHEN¨ Institut fur¨ Informatik Lehrstuhl Informatik II Interprocedural Analysis of Low-Level Code Andrea Flexeder Vollstandiger¨ Abdruck der von der Fakultat¨ fur¨ Informatik der Technischen Universitat¨ Munchen¨ zur Erlangung des akademischen Grades eines Doktors der Naturwissenschaften (Dr. rer. nat.) genehmigten Dissertation. Vorsitzender: Univ.-Prof. Dr. H. M. Gerndt Prufer¨ der Dissertation: 1. Univ.-Prof. Dr. H. Seidl 2. Dr. A. King, University of Kent at Canterbury / UK Die Dissertation wurde am 14.12.2010 bei der Technischen Universitat¨ Munchen¨ eingereicht und durch die Fakultat¨ fur¨ Informatik am 9.6.2011 angenommen. ii Contents 1 Analysis of Low-Level Code 1 1.1 Source versus Binary . 1 1.2 Application Areas . 6 1.3 Executable and Linkable Format (ELF) .................. 12 1.4 Application Binary Interface (ABI)..................... 18 1.5 Assumptions . 24 1.6 Contributions . 24 2 Control Flow Reconstruction 27 2.1 The Concrete Semantics . 31 2.2 Interprocedural Control Flow Reconstruction . 33 2.3 Practical Issues . 39 2.4 Implementation . 43 2.5 Programming Model . 44 3 Classification of Memory Locations 49 3.1 Semantics . 51 3.2 Interprocedural Variable Differences . 58 3.3 Application to Assembly Analysis . 73 4 Reasoning about Array Index Expressions 81 4.1 Linear Two-Variable Equalities . 81 4.2 Application to Assembly Analysis . 88 4.3 Register Coalescing and Locking . 89 5 Tools 91 5.1 Combination of Abstract Domains . 91 5.2 VoTUM . 96 6 Side-Effect Analysis 101 6.1 Semantics . 105 6.2 Analysis of Side-Effects . 108 6.3 Enhancements . 115 6.4 Experimental Results . 118 iii iv CONTENTS 7 Exploiting Alignment for WCET and Data Structures 123 7.1 Alignment Analysis .
    [Show full text]
  • Mac OS X ABI Function Call Guide
    Mac OS X ABI Function Call Guide 2005-12-06 PowerPC and and the PowerPC logo are Apple Computer, Inc. trademarks of International Business © 2005 Apple Computer, Inc. Machines Corporation, used under license All rights reserved. therefrom. Simultaneously published in the United No part of this publication may be States and Canada. reproduced, stored in a retrieval system, or Even though Apple has reviewed this document, transmitted, in any form or by any means, APPLE MAKES NO WARRANTY OR mechanical, electronic, photocopying, REPRESENTATION, EITHER EXPRESS OR IMPLIED, WITH RESPECT TO THIS recording, or otherwise, without prior DOCUMENT, ITS QUALITY, ACCURACY, written permission of Apple Computer, Inc., MERCHANTABILITY, OR FITNESS FOR A with the following exceptions: Any person PARTICULAR PURPOSE. AS A RESULT, THIS DOCUMENT IS PROVIDED “AS IS,” AND is hereby authorized to store documentation YOU, THE READER, ARE ASSUMING THE on a single computer for personal use only ENTIRE RISK AS TO ITS QUALITY AND ACCURACY. and to print copies of documentation for IN NO EVENT WILL APPLE BE LIABLE FOR personal use provided that the DIRECT, INDIRECT, SPECIAL, INCIDENTAL, documentation contains Apple’s copyright OR CONSEQUENTIAL DAMAGES notice. RESULTING FROM ANY DEFECT OR INACCURACY IN THIS DOCUMENT, even if The Apple logo is a trademark of Apple advised of the possibility of such damages. Computer, Inc. THE WARRANTY AND REMEDIES SET FORTH ABOVE ARE EXCLUSIVE AND IN Use of the “keyboard” Apple logo LIEU OF ALL OTHERS, ORAL OR WRITTEN, EXPRESS OR IMPLIED. No Apple dealer, agent, (Option-Shift-K) for commercial purposes or employee is authorized to make any without the prior written consent of Apple modification, extension, or addition to this may constitute trademark infringement and warranty.
    [Show full text]
  • Alignment in C Seminar “Effiziente Programmierung in C”
    Introduction Data Structure Alignment Heap Alignment Stack Alignment Summary Alignment in C Seminar “Effiziente Programmierung in C" Sven-Hendrik Haase Universit¨atHamburg, Fakult¨atf¨urInformatik 2014-01-09 Sven-Hendrik Haase Seminar \Effiziente Programmierung in C" 1/ 32 Introduction Data Structure Alignment Heap Alignment Stack Alignment Summary Outline Introduction Guiding Questions of This Presentation Memory Addressing Alignment 101 Consquences of Misalignment Different Types of Alignment Data Structure Alignment Structs and Stuff Padding in the Real World Performance Implications SSE Heap Alignment Introduction Example Use Cases Stack Alignment Introduction The Problem Use Cases Summary TL;DR Resources Sven-Hendrik Haase Seminar \Effiziente Programmierung in C" 2/ 32 • What is data alignment? • What is heap alignment? • What is stack alignment? • How does it work in C? • Do we need to care abouy any of these? Introduction Data Structure Alignment Heap Alignment Stack Alignment Summary Introduction Guiding Questions of This Presentation • Which types of alignment exist in C? Sven-Hendrik Haase Seminar \Effiziente Programmierung in C" 3/ 32 • What is heap alignment? • What is stack alignment? • How does it work in C? • Do we need to care abouy any of these? Introduction Data Structure Alignment Heap Alignment Stack Alignment Summary Introduction Guiding Questions of This Presentation • Which types of alignment exist in C? • What is data alignment? Sven-Hendrik Haase Seminar \Effiziente Programmierung in C" 3/ 32 • What is stack alignment? •
    [Show full text]
  • Compiler Construction
    Compiler construction PDF generated using the open source mwlib toolkit. See http://code.pediapress.com/ for more information. PDF generated at: Sat, 10 Dec 2011 02:23:02 UTC Contents Articles Introduction 1 Compiler construction 1 Compiler 2 Interpreter 10 History of compiler writing 14 Lexical analysis 22 Lexical analysis 22 Regular expression 26 Regular expression examples 37 Finite-state machine 41 Preprocessor 51 Syntactic analysis 54 Parsing 54 Lookahead 58 Symbol table 61 Abstract syntax 63 Abstract syntax tree 64 Context-free grammar 65 Terminal and nonterminal symbols 77 Left recursion 79 Backus–Naur Form 83 Extended Backus–Naur Form 86 TBNF 91 Top-down parsing 91 Recursive descent parser 93 Tail recursive parser 98 Parsing expression grammar 100 LL parser 106 LR parser 114 Parsing table 123 Simple LR parser 125 Canonical LR parser 127 GLR parser 129 LALR parser 130 Recursive ascent parser 133 Parser combinator 140 Bottom-up parsing 143 Chomsky normal form 148 CYK algorithm 150 Simple precedence grammar 153 Simple precedence parser 154 Operator-precedence grammar 156 Operator-precedence parser 159 Shunting-yard algorithm 163 Chart parser 173 Earley parser 174 The lexer hack 178 Scannerless parsing 180 Semantic analysis 182 Attribute grammar 182 L-attributed grammar 184 LR-attributed grammar 185 S-attributed grammar 185 ECLR-attributed grammar 186 Intermediate language 186 Control flow graph 188 Basic block 190 Call graph 192 Data-flow analysis 195 Use-define chain 201 Live variable analysis 204 Reaching definition 206 Three address
    [Show full text]
  • An Introduction to Reverse Engineering for Beginners
    An Introduction To Reverse Engineering for Beginners Dennis Yurichev <[email protected]> cbnd ○c 2013, Dennis Yurichev. This work is licensed under the Creative Commons Attribution-NonCommercial-NoDerivs 3.0 Unported License. To view a copy of this license, visit http://creativecommons.org/licenses/by-nc-nd/3.0/. Text version (November 23, 2013). There is probably a newer version of this text, and also Russian language version also accessible at http://yurichev.com/RE-book.html You may also subscribe to my twitter, to get information about updates of this text, etc: @yurichev, or to subscribe to mailing list. CONTENTS CONTENTS Contents Preface vi 0.1 Mini-FAQ1 ................................................. vi 0.2 About the author............................................. vi 0.3 Thanks.................................................. vi 0.4 Donate.................................................. vii 1 Compiler’s patterns 1 1.1 Hello, world!...............................................1 1.1.1 x86................................................1 1.1.2 ARM................................................4 1.2 Stack...................................................8 1.2.1 What is the stack used for?...................................9 1.3 printf() with several arguments.................................... 13 1.3.1 x86................................................ 13 1.3.2 ARM: 3 printf() arguments.................................. 14 1.3.3 ARM: 8 printf() arguments.................................. 15 1.3.4 By the way...........................................
    [Show full text]
  • Unstructured Computations on Emerging Architectures
    Unstructured Computations on Emerging Architectures Dissertation by Mohammed A. Al Farhan In Partial Fulfillment of the Requirements For the Degree of Doctor of Philosophy King Abdullah University of Science and Technology Thuwal, Kingdom of Saudi Arabia May 2019 2 EXAMINATION COMMITTEE PAGE The dissertation of M. A. Al Farhan is approved by the examination committee Dissertation Committee: David E. Keyes, Chair Professor, King Abdullah University of Science and Technology Edmond Chow Associate Professor, Georgia Institute of Technology Mikhail Moshkov Professor, King Abdullah University of Science and Technology Markus Hadwiger Associate Professor, King Abdullah University of Science and Technology Hakan Bagci Associate Professor, King Abdullah University of Science and Technology 3 ©May 2019 Mohammed A. Al Farhan All Rights Reserved 4 ABSTRACT Unstructured Computations on Emerging Architectures Mohammed A. Al Farhan his dissertation describes detailed performance engineering and optimization Tof an unstructured computational aerodynamics software system with irregu- lar memory accesses on various multi- and many-core emerging high performance computing scalable architectures, which are expected to be the building blocks of energy-austere exascale systems, and on which algorithmic- and architecture-oriented optimizations are essential for achieving worthy performance. We investigate several state-of-the-practice shared-memory optimization techniques applied to key kernels for the important problem class of unstructured meshes. We illustrate
    [Show full text]
  • AUTHOR PUB TYPE Needs of Potential Users, System
    DOCUMENT RESUME ED 355 773 FL 021 026 AUTHOR Mao, Tang TITLE Courseware Authoring and Delivering System for Chinese Language Instruction. Final Report. INSTITUTION Comptek Co., Springfield, NJ. SPONE AGENCY Department of Education, Washington, DC. PUB DATE 5 Feb 85 CONTRACT 300-84-0180 NOTE 41p. PUB TYPE Reports Descriptive (141) EDRS PRICE MF01/PCO2 Plus Postage. DESCRIPTORS *Authoring Aids (Programing); Chinese; *Computer Assisted Instruction; Computer Software Development; *Courseware; *Programing; Research Projects; Second Language Instruction; *Second Languages; *Uncommonly Taught Languages IDENTIFIERS *Unix Operating System ABSTRACT A study investigated technical methods for simplifying and improving the creation of software for teaching uncommonly taught languages such as Chinese. Research consisted of assessment of existing authoring systems, domestic and overseas, available hardware, peripherals, and software packages that could be integrated into this projectThen some features of an authoring system, were applied to a personal computer using available hardware and software. Among the findings were that interactive computer-assisted instruction (CAI) applications have large storage requirements; choice of hardware and peripherals require trade-offs in cost, speed, and adequacy; a courseware author is needed; the program can be designed to incorporate a third language; data compatibility is the primary difficulty in transferring information from one computer type to another; and the Unix system is inadequate for this application. It is concluded that an interactive CAI system for multiple languages with graphics, text, and voice capability at reasonable cost can be developed if special care is given to the needs of potential users, system development, and integration of hardware and software. The final product of the research is a detailed design for a multi-language, multimedia courseware authoring and delivery system.
    [Show full text]
  • Program Optimization
    Program optimization Alexey A. Romanenko [email protected] What this course about? In this course you will learn: ● Types of program optimization ● Approaches for program optimization ● Profiling tools Optimization Optimization consists of analyzing and tuning software code to make it perform faster and more efficiently on Intel processor architecture. www.intel.com/products/glossary/body.htm On ANY architecture! Types of optimization ● Performance ● Memory ● Scalability ● Communications ● etc. ● We address mainly to performance optimization Statements ● Donald Knuth – Premature optimization is the root of all evil ● If your code already works, optimizing it is a sure way to introduce new, and possibly subtle, bugs ● Optimization tends to make code harder to understand and maintain ● Some of the optimization techniques increase speed by reducing the extensibility of the code ● A lot of time can be spent optimizing, with little gain in performance, and can result in obfuscated code ● If you're overly obsessed with optimizing code, people will call you a nerd behind your back Statements ● Faster program, more memory required. ● Making more faster program required more time for optimization. ● Optimizing code for one platform may actually make it worse on another platform BUT!!!!! ● Michael Abrash (Quake, Unreal Tournament, Space Strike, Half-life, etc.): – Performance must always be measured – Any optimization, user can feel, should be done. Does compilers make an optimization? ● They does. But – Compiler do not have database for algorithms
    [Show full text]
  • A Guide to Vectorization with Intel® C++ Compilers
    A Guide to Vectorization with Intel® C++ Compilers Contents 1. Introduction .............................................................................................................................................. 4 2. What is Vectorization in the Intel Compiler? ............................................................................... 4 2.1 When does the compiler try to vectorize? ............................................................................... 6 2.2 How do I know whether a loop was vectorized? .................................................................. 6 2.3 What difference does it make?...................................................................................................... 7 3. What sort of loops can be vectorized? ............................................................................................ 7 4. Obstacles to vectorization ................................................................................................................ 10 4.1 Non-contiguous Memory Accesses ............................................................................................ 10 4.2 Data Dependencies ........................................................................................................................... 11 5. Guidelines for writing vectorizable code ................................................................................... 13 5.1 General Guidelines ....................................................................................................................
    [Show full text]
  • Gdura, Youssef Omran (2012) a New Parallelisation Technique for Heterogeneous Cpus
    Gdura, Youssef Omran (2012) A new parallelisation technique for heterogeneous CPUs. PhD thesis http://theses.gla.ac.uk/3406/ Copyright and moral rights for this thesis are retained by the author A copy can be downloaded for personal non-commercial research or study, without prior permission or charge This thesis cannot be reproduced or quoted extensively from without first obtaining permission in writing from the Author The content must not be changed in any way or sold commercially in any format or medium without the formal permission of the Author When referring to this work, full bibliographic details including the author, title, awarding institution and date of the thesis must be given Glasgow Theses Service http://theses.gla.ac.uk/ [email protected] . A New Parallelisation Technique for Heterogeneous CPUs by Youssef Omran Gdura A Thesis presented for the degree of Doctor of Philosophy to the School of Computing Science, College of Science and Engineering, University of Glasgow May 2012 . Copyright © Youssef Omran Gdura, 2012. Abstract Parallelization has moved in recent years into the mainstream compilers, and the demand for parallelizing tools that can do a better job of automatic parallelization is higher than ever. During the last decade considerable attention has been focused on developing pro- gramming tools that support both explicit and implicit parallelism to keep up with the power of the new multiple core technology. Yet the success to develop automatic paral- lelising compilers has been limited mainly due to the complexity of the analytic process required to exploit available parallelism and manage other parallelisation measures such as data partitioning, alignment and synchronization.
    [Show full text]