An Optimized Mapreduce Framework for AMD Gpus

An Optimized Mapreduce Framework for AMD Gpus

StreamMR: An Optimized MapReduce Framework for AMD GPUs Marwa Elteir∗y, Heshan Liny, Wu-chun Fengy, and Tom Scoglandy ∗City of Scientific Researches and Technology Applications, Egypt yDepartment of Computer Science, Virginia Tech Emails: fmaelteir, hlin2, feng, [email protected] Abstract—MapReduce is a programming model from Google CompletePath, as opposed to the normal memory-access that facilitates parallel processing on a cluster of thousands of path, i.e., FastPath, even if the atomic operation is not commodity computers. The success of MapReduce in cluster executed.1 Our results show that for certain applications, the environments has motivated several studies of implementing MapReduce on a graphics processing unit (GPU), but generally atomic-based implementation of MapReduce can introduce focusing on the NVIDIA GPU. severe performance degradation, e.g., a 28-fold slowdown. Our investigation reveals that the design and mapping of the Although Mars [4] is an atomic-free implementation of MapReduce framework needs to be revisited for AMD GPUs MapReduce on GPUs, it has several disadvantages. First, due to their notable architectural differences from NVIDIA Mars incurs expensive preprocessing phases (i.e., redundant GPUs. For instance, current state-of-the-art MapReduce imple- mentations employ atomic operations to coordinate the execution counting of output records and prefix summing) in order to of different threads. However, atomic operations can implicitly coordinate result writing of different threads. Second, Mars cause inefficient memory access, and in turn, severely impact sorts the keys to group intermediate results generated by the performance. In this paper, we propose StreamMR, an OpenCL map function, which has been found inefficient [5]. MapReduce framework optimized for AMD GPUs. With efficient In this paper, we propose StreamMR, an OpenCL atomic-free algorithms for output handling and intermediate re- sult shuffling, StreamMR is superior to atomic-based MapReduce MapReduce framework optimized for AMD GPUs. The designs and can outperform existing atomic-free MapReduce design and mapping of StreamMR provides efficient atomic- implementations by nearly five-fold on an AMD Radeon HD 5870. free algorithms for coordinating output from different threads Index Terms—atomics, parallel computing, AMD GPU, as well as storing and retrieving intermediate results via hash GPGPU, MapReduce, Mars, MapCG, OpenCL tables. StreamMR also includes efficient support of combiner functions, a feature widely used in cluster MapReduce I. INTRODUCTION implementations but not well explored in previous GPU While graphics processing units (GPUs) were originally de- MapReduce implementations. The performance results of signed to accelerate data-parallel, graphics-based applications, three real-world applications show that StreamMR is superior the introduction of programming models such as CUDA [15], to atomic-based MapReduce designs and can outperform an Brook+ [2], and OpenCL [10] has made general-purpose existing atomic-free MapReduce framework (i.e., Mars) by computing on the GPU (i.e., GPGPU) a reality. Although nearly five-fold on AMD GPUs. GPUs have the potential of delivering astounding raw performance via the above programming models, developing II. BACKGROUND and optimizing programs on GPUs requires intimate A. AMD GPU Architecture knowledge of the architectural details, and thus, is nontrivial. An AMD GPU consists of multiple SIMD units named High-level programming models such as MapReduce play compute units, and each compute unit consists of several an essential role in hiding architectural details of parallel cores called stream cores. Each stream core is a VLIW computing platforms from programmers. MapReduce, pro- processor containing five processing elements, with one of posed by Google [7], seeks to simplify parallel programming them capable of performing transcendental operations like on large-scale clusters of computers. With MapReduce, users sine, cosine, and logarithm. Each compute unit also contains only need to write a map function and a reduce function, a branch execution unit that handles branch instructions. All and the parallel execution and fault tolerance is handled stream cores on a compute unit execute the same instruction by the runtime framework. The success of MapReduce on sequence in a lock-step fashion. cluster environments has also motivated studies of porting There is a low-latency, on-chip memory region shared by MapReduce on other parallel platforms including multicore all stream cores on a compute unit named LDS. Each LDS is systems [6], [16], Cell [12], and GPUs [16], [5], [9], [18]. connected to L1 cache as shown in Figure 1. Several compute However, existing MapReduce implementations on GPUs units then share one L2 cache that is connected to the global focus on NVIDIA GPUs. The design and optimization memory through a memory controller. The global memory is techniques in these implementations may not be applicable to a high-latency, off-chip memory that is shared by all compute AMD GPUs, which have a considerably different architecture units. Host CPU transfers the data to global memory through than NVIDIA ones. For instance, state-of-the-art MapReduce PCIe bus. In addition to the local and global memory, there implementations on NVIDIA GPUs [5], [9] rely on atomic are two special types of memories that are also shared by all operations to coordinate execution of different threads. But as compute units. Image memory is a high-bandwidth memory the AMD OpenCL programming guide notes [3], including an atomic operation in a GPU kernel may cause all memory 1The details of both CompletePath and FastPath will be discussed in accesses to follow a much slower memory-access path, i.e., Section II. region whose reads are cached through L1 and L2 caches. D. MapReduce Programming Model Constant memory is a memory region storing data that are MapReduce is a high-level programming model aims at allocated/initialized by the host and not changed during the facilitating parallel programming by masking the details of kernel execution. Access to constant memory is also cached. the underling architecture. Programmers need only to write their applications as two functions: the map function and the B. Memory Paths reduce function. All of the input and outputs are represented as key/value pairs. Implementing a MapReduce framework As shown in Figure 1, ATI Radeon HD 5000 series GPUs involves implementing three phases: the map phase, the group have two independent paths for memory access: FastPath phase, and the reduce phase. Specifically, the MapReduce and CompletePath [3]. The bandwidth of the FastPath is framework first partitions the input dataset among the partic- significantly higher than the CompletePath. Loads and stores ipating parties (e.g. threads). Each party then applies the map of data whose size is multiple of 32 bits are executed through function to its assigned portion and writes the intermediate FastPath, whereas advanced operations like atomics and sub- output (map phase). The framework groups all of the interme- 32 bit data transfers are executed through the CompletePath. diate outputs by their keys (group phase). Finally, one or more Executing a memory load access through FastPath is keys of the grouped intermediate outputs are assigned to each performed by a single vertex fetch (vfetch) instruction. In partition party, which will carry out the reducing function contrast, a memory load through the CompletePath requires and write out the result key/value pairs (reduce phase). a multi-phase operation and thus can be several folds slower III. RELATED WORK according to the AMD OpenCL programming guide [3]. On AMD GPUs, the selection of the memory path is done Many research efforts have been done to enhance the automatically by the compiler. The current OpenCL compiler MapReduce framework [11], [19], [17], [14], [13] in cluster maps all kernel data into a single unordered access view. environments. Valvag et al. developed a high-level declarative Consequently, including a single atomic operation in a programming model and its underlying runtime, Oivos, which kernel may force all memory loads and stores to follow aims at handling applications that require running several the CompletePath instead of the FastPath, which can in MapReduce jobs [19]. Zahria et al. [14] on the other side turn cause severe performance degradation of an application proposed a speculative task scheduling named LATE (Longest as discovered by our previous study [8]. Note that atomic Approximate Time to End) to cope with several limitations operations on variables stored in the local memory does not of the original Hadoop’s scheduler in heterogeneous impact the selection of memory path. environments such as Amazon EC2[1]. Moreover, Elteir et al. [13] recently enhanced MapReduce framework to support asynchronous data processing. Instead of having SIMD Engine barrier synchronization between map and reduce phases, they LDS, Registers propose interleaving both phases, and start the reduce phase L 1 Cache as soon as a specific number of map tasks are finished. Mars [4] is the first MapReduce implementation on GPUs. Compute Unit to Memory X-bar One of the main challenges of implementing MapReduce on GPUs is to safely write the output to a global buffer without conflicting with output from other threads. Mars addresses L2 Cache Write Cache this by calculating the exact write location of each thread. Specifically, it executes two preprocessing kernels before

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    8 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us