Index-Based Search Techniques for Visualization and Data Analysis Algorithms on Many-Core Systems

Total Page:16

File Type:pdf, Size:1020Kb

Index-Based Search Techniques for Visualization and Data Analysis Algorithms on Many-Core Systems INDEX-BASED SEARCH TECHNIQUES FOR VISUALIZATION AND DATA ANALYSIS ALGORITHMS ON MANY-CORE SYSTEMS by BRENTON JOHN LESSLEY A DISSERTATION Presented to the Department of Computer and Information Science and the Graduate School of the University of Oregon in partial fulfillment of the requirements for the degree of Doctor of Philosophy June 2019 DISSERTATION APPROVAL PAGE Student: Brenton John Lessley Title: Index-Based Search Techniques for Visualization and Data Analysis Algorithms on Many-Core Systems This dissertation has been accepted and approved in partial fulfillment of the requirements for the Doctor of Philosophy degree in the Department of Computer and Information Science by: Hank Childs Chair Boyana Norris Core Member Christopher Wilson Core Member Eric Torrence Institutional Representative and Janet Woodruff-Borden Vice Provost and Dean of the Graduate School Original approval signatures are on file with the University of Oregon Graduate School. Degree awarded June 2019 ii ⃝c 2019 Brenton John Lessley All rights reserved. iii DISSERTATION ABSTRACT Brenton John Lessley Doctor of Philosophy Department of Computer and Information Science June 2019 Title: Index-Based Search Techniques for Visualization and Data Analysis Algorithms on Many-Core Systems Sorting and hashing are canonical index-based methods to perform searching, and are often sub-routines in many visualization and analysis algorithms. With the emergence of many-core architectures, these algorithms must be rethought to exploit the increased available thread-level parallelism and data-parallelism. Data-parallel primitives (DPP) provide an efficient way to design an algorithm for scalable, platform-portable parallelism. This dissertation considers the following question: What are the best index- based search techniques for visualization and analysis algorithms on diverse many-core systems? To answer this question, we develop new DPP-based techniques, and evaluate their performance against existing techniques for data-intensive visualization and analysis algorithms across different many-core platforms. Then, we synthesize our findings into a collection of best practices and recommended usage. As a result of these efforts, we were able to conclude that our techniques demonstrate viability and leading platform- portable performance for several different search-based use cases. This dissertation is a culmination of previously-published co-authored material. iv CURRICULUM VITAE NAME OF AUTHOR: Brenton John Lessley DEGREES AWARDED: Ph.D. in Computer Science University of Oregon, 2019 M.S. in Computer Science University of Washington Tacoma, 2011 B.S. in Business Administration University of Washington Tacoma, 2009 AREAS OF SPECIAL INTEREST: Data-Parallel Algorithms Parallel Data Structures High-Performance Computing Scientific Visualization PROFESSIONAL EXPERIENCE: Graduate Research & Teaching Fellow University of Oregon, 2013 - 2019 Graduate Student Researcher Lawrence Berkeley Nat’l Lab, Summer 2017 Graduate Student Researcher Lawrence Berkeley Nat’l Lab, Summer 2016 Student Developer Google Summer of Code, Summer 2015 PUBLICATIONS: Brenton Lessley, Samuel Li, and Hank Childs (2019). HashFight: A Platform- Portable Hash Table for Multi-Core and Many-Core Architectures (in preparation). Brenton Lessley and Hank Childs (2019). Data-Parallel Hashing Techniques for GPU Architectures (in submission). v Brenton Lessley, Talita Perciano, Colleen Heinemann, David Camp, Hank Childs, and E. Wes Bethel (2018). DPP-PMRF: Rethinking Optimization for a Probabilistic Graphical Model Using Data-Parallel Primitives. Proceedings of the IEEE Symposium on Large Data Visualization and Analysis (LDAV), Berlin, Germany, pp. 1–11. Brenton Lessley, Kenneth Moreland, Matthew Larsen and Hank Childs (2017). Techniques for Data-Parallel Searching for Duplicate Elements. Proceedings of the IEEE Symposium on Large Data Visualization and Analysis (LDAV), Phoenix, AZ, pp. 1–5. Brenton Lessley, Talita Perciano, Manish Mathai, Hank Childs and E. Wes Bethel (2017). Maximal Clique Enumeration with Data-Parallel Primitives. Proceedings of the IEEE Symposium on Large Data Visualization and Analysis (LDAV), Phoenix, AZ, pp. 16–25. Brenton Lessley, Roba Binyahib, Robert Maynard, and Hank Childs (2016). External Facelist Calculation with Data-Parallel Primitives. Proceedings of EuroGraphics Symposium on Parallel Graphics and Visualization (EGPGV), Groningen, The Netherlands, pp. 10–20. Sergio Davalos, Altaf Merchant, Gregory Rose, Brenton Lessley, and Ankur Teredesai (2015). ’The Good Old Days’: An Examination of Nostalgia in Facebook Posts. International Journal of Human-Computer Studies, vol. 83, pp. 83–93. Daniel Lowd, Brenton Lessley, and Mino De Raj (2014). Towards Adversarial Reasoning in Statistical Relational Domains. AAAI-14 Workshop on Statistical Relational AI (Star AI 2014), Quebec City, Canada. Brenton Lessley, Matthew Alden, Daniel Bryan, and Arindam Tripathy (2012). Detection of Financial Statement Fraud Using Evolutionary Algorithms. Journal of Emerging Technologies in Accounting, vol. 9, no. 1, pp. 71–94. vi ACKNOWLEDGEMENTS The path towards completing my Ph.D. has been quite an adventure, always seeing the light at the end of the tunnel, yet always one more challenging task to complete. In short, the past six years have more or less been a mind-training game that teaches me how to stay focused on a goal and work towards it tirelessly until completion, despite countless thoughts of giving up. To reach this point took more self-motivation than I ever knew I had and even more motivation and encouragement from those around me. Throughout the Ph.D. process, my parents have provided the best support, advice, and encouragement that I could have asked for. They have been my number one cheering squad and, with great humor, make sure that I always remain optimistic, from one paper or milestone to the next. I cannot thank them enough. The same thank you goes to my very good friend, Agrima, who has patiently heard every high note and low note from me while completing this long process. I know she sacrificed a lot of fun to join in on my many work sessions or postpone a trip due to a looming milestone. I truly appreciate her patience and sense of humor, and can’t thank her enough. If it wasn’t for my advisor, Hank Childs, I likely would not have even completed the PhD degree. From the day I joined the CDUX research group and planned out a road map towards completing the PhD, Hank has helped me become a self-sufficient researcher and provided countless opportunities to hone my area of expertise. I thank him for all the guidance and support he has provided me, particularly in the past year. I am proud of all the work that we have completed and published during the four years under his advisement. vii During my time at the University of Oregon, I have made some of my best friends, who provided so many nice memories and made the hard work much more enjoyable, even if it was as simple as going to a basketball or football game after completing some goal, holding a house party, or going to Starbucks for a coding session. You know who you are—thank you, you are the best. I was fortunate to have many awesome officemates during my six years in Deschutes Hall, all of whom provided plenty of great humor, language learning, and a benchmark for working harder. I can confidently say that our whiteboards were rarely empty. To Ali, Pedram, Azaad, Soheil, Sam, Vince, Stephanie, Abhishek, Manish, Garrett, Sudhanshu, and Steve, it was a great pleasure. This thank you also extends to my fellow labmates and friends in CDUX. From the group parties and practice presentations to random office conversations, your sense of humor and friendship were awesome. While most of my time was spent at the University of Oregon, I also spent two great summers at the Lawrence Berkeley National Lab with the Data Analytics & Visualization Group. The research projects and papers I completed during this time really boosted my confidence as a researcher and led to a pipeline of exciting work. I was lucky to have Wes Bethel and Talita Perciano as my mentors, and I appreciated the discussions and collaboration with David Camp, Colleen Heinemann, Dmitriy Morozov, Sugeerth Murugesan, Dani Ushizima, Burlen Loring, and Hari Krishnan. Finally, I would like to thank Boyana Norris, Chris Wilson, and Eric Torrence for their willingness to serve on my PhD committee and provide great feedback throughout the dissertation phase. viii TABLE OF CONTENTS Chapter Page I. INTRODUCTION ................................ 1 1.1. Dissertation Outline .............................. 4 1.2. Co-Authored Material ............................. 5 I Techniques 7 II. BACKGROUND ................................. 9 2.1. Parallel Computing & Data-Parallelism .................... 9 2.2. Data Parallel Primitives ............................ 12 2.3. General-Purpose GPU Computing ...................... 15 2.3.1. SIMT Architecture .......................... 17 2.3.2. Optimal Performance Criteria .................... 19 2.4. Index-Based Searching ............................ 22 2.4.1. Searching Via Sorting ......................... 24 2.4.2. Searching Via Hashing ........................ 25 2.5. Data-Parallel Hashing Techniques ...................... 30 2.5.1. Open-addressing Probing ....................... 31 2.5.1.1. Linear Probing-based Hashing . 32 2.5.1.2. Cuckoo-based Hashing ................... 33 2.5.1.3. Double Hashing ...................... 38 2.5.1.4. Robin Hood-based Hashing .
Recommended publications
  • Workload Management and Application Placement for the Cray Linux Environment™
    TMTM Workload Management and Application Placement for the Cray Linux Environment™ S–2496–4101 © 2010–2012 Cray Inc. All Rights Reserved. This document or parts thereof may not be reproduced in any form unless permitted by contract or by written permission of Cray Inc. U.S. GOVERNMENT RESTRICTED RIGHTS NOTICE The Computer Software is delivered as "Commercial Computer Software" as defined in DFARS 48 CFR 252.227-7014. All Computer Software and Computer Software Documentation acquired by or for the U.S. Government is provided with Restricted Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7014, as applicable. Technical Data acquired by or for the U.S. Government, if any, is provided with Limited Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7013, as applicable. Cray and Sonexion are federally registered trademarks and Active Manager, Cascade, Cray Apprentice2, Cray Apprentice2 Desktop, Cray C++ Compiling System, Cray CX, Cray CX1, Cray CX1-iWS, Cray CX1-LC, Cray CX1000, Cray CX1000-C, Cray CX1000-G, Cray CX1000-S, Cray CX1000-SC, Cray CX1000-SM, Cray CX1000-HN, Cray Fortran Compiler, Cray Linux Environment, Cray SHMEM, Cray X1, Cray X1E, Cray X2, Cray XD1, Cray XE, Cray XEm, Cray XE5, Cray XE5m, Cray XE6, Cray XE6m, Cray XK6, Cray XK6m, Cray XMT, Cray XR1, Cray XT, Cray XTm, Cray XT3, Cray XT4, Cray XT5, Cray XT5h, Cray XT5m, Cray XT6, Cray XT6m, CrayDoc, CrayPort, CRInform, ECOphlex, LibSci, NodeKARE, RapidArray, The Way to Better Science, Threadstorm, uRiKA, UNICOS/lc, and YarcData are trademarks of Cray Inc.
    [Show full text]
  • Introducing the Cray XMT
    Introducing the Cray XMT Petr Konecny Cray Inc. 411 First Avenue South Seattle, Washington 98104 [email protected] May 5th 2007 Abstract Cray recently introduced the Cray XMT system, which builds on the parallel architecture of the Cray XT3 system and the multithreaded architecture of the Cray MTA systems with a Cray-designed multithreaded processor. This paper highlights not only the details of the architecture, but also the application areas it is best suited for. 1 Introduction past two decades the processor speeds have increased several orders of magnitude while memory speeds Cray XMT is the third generation multithreaded ar- increased only slightly. Despite this disparity the chitecture built by Cray Inc. The two previous gen- performance of most applications continued to grow. erations, the MTA-1 and MTA-2 systems [2], were This has been achieved in large part by exploiting lo- fully custom systems that were expensive to man- cality in memory access patterns and by introducing ufacture and support. Developed under code name hierarchical memory systems. The benefit of these Eldorado [4], the Cray XMT uses many parts built multilevel caches is twofold: they lower average la- for other commercial system, thereby, significantly tency of memory operations and they amortize com- lowering system costs while maintaining the sim- munication overheads by transferring larger blocks ple programming model of the two previous genera- of data. tions. Reusing commodity components significantly The presence of multiple levels of caches forces improves price/performance ratio over the two pre- programmers to write code that exploits them, if vious generations. they want to achieve good performance of their ap- Cray XMT leverages the development of the Red plication.
    [Show full text]
  • The Gemini Network
    The Gemini Network Rev 1.1 Cray Inc. © 2010 Cray Inc. All Rights Reserved. Unpublished Proprietary Information. This unpublished work is protected by trade secret, copyright and other laws. Except as permitted by contract or express written permission of Cray Inc., no part of this work or its content may be used, reproduced or disclosed in any form. Technical Data acquired by or for the U.S. Government, if any, is provided with Limited Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7013, as applicable. Autotasking, Cray, Cray Channels, Cray Y-MP, UNICOS and UNICOS/mk are federally registered trademarks and Active Manager, CCI, CCMT, CF77, CF90, CFT, CFT2, CFT77, ConCurrent Maintenance Tools, COS, Cray Ada, Cray Animation Theater, Cray APP, Cray Apprentice2, Cray C90, Cray C90D, Cray C++ Compiling System, Cray CF90, Cray EL, Cray Fortran Compiler, Cray J90, Cray J90se, Cray J916, Cray J932, Cray MTA, Cray MTA-2, Cray MTX, Cray NQS, Cray Research, Cray SeaStar, Cray SeaStar2, Cray SeaStar2+, Cray SHMEM, Cray S-MP, Cray SSD-T90, Cray SuperCluster, Cray SV1, Cray SV1ex, Cray SX-5, Cray SX-6, Cray T90, Cray T916, Cray T932, Cray T3D, Cray T3D MC, Cray T3D MCA, Cray T3D SC, Cray T3E, Cray Threadstorm, Cray UNICOS, Cray X1, Cray X1E, Cray X2, Cray XD1, Cray X-MP, Cray XMS, Cray XMT, Cray XR1, Cray XT, Cray XT3, Cray XT4, Cray XT5, Cray XT5h, Cray Y-MP EL, Cray-1, Cray-2, Cray-3, CrayDoc, CrayLink, Cray-MP, CrayPacs, CrayPat, CrayPort, Cray/REELlibrarian, CraySoft, CrayTutor, CRInform, CRI/TurboKiva, CSIM, CVT, Delivering the power…, Dgauss, Docview, EMDS, GigaRing, HEXAR, HSX, IOS, ISP/Superlink, LibSci, MPP Apprentice, ND Series Network Disk Array, Network Queuing Environment, Network Queuing Tools, OLNET, RapidArray, RQS, SEGLDR, SMARTE, SSD, SUPERLINK, System Maintenance and Remote Testing Environment, Trusted UNICOS, TurboKiva, UNICOS MAX, UNICOS/lc, and UNICOS/mp are trademarks of Cray Inc.
    [Show full text]
  • Overview of the Next Generation Cray XMT
    Overview of the Next Generation Cray XMT Andrew Kopser and Dennis Vollrath, Cray Inc. ABSTRACT: The next generation of Cray’s multithreaded supercomputer is architecturally very similar to the Cray XMT, but the memory system has been improved significantly and other features have been added to improve reliability, productivity, and performance. This paper begins with a review of the Cray XMT followed by an overview of the next generation Cray XMT implementation. Finally, the new features of the next generation XMT are described in more detail and preliminary performance results are given. KEYWORDS: Cray, XMT, multithreaded, multithreading, architecture 1. Introduction The primary objective when designing the next Cray XMT generation Cray XMT was to increase the DIMM The Cray XMT is a scalable multithreaded computer bandwidth and capacity. It is built using the Cray XT5 built with the Cray Threadstorm processor. This custom infrastructure with DDR2 technology. This change processor is designed to exploit parallelism that is only increases each node’s memory capacity by a factor of available through its unique ability to rapidly context- eight and DIMM bandwidth by a factor of three. switch among many independent hardware execution streams. The Cray XMT excels at large graph problems Reliability and productivity were important goals for that are poorly served by conventional cache-based the next generation Cray XMT as well. A source of microprocessors. frustration, especially for new users of the current Cray XMT, is the issue of hot spots. Since all memory is The Cray XMT is a shared memory machine. The programming model assumes a flat, globally accessible, globally accessible and the application has access to shared address space.
    [Show full text]
  • Pubtex Output 2011.12.12:1229
    TM Using Cray Performance Analysis Tools S–2376–53 © 2006–2011 Cray Inc. All Rights Reserved. This document or parts thereof may not be reproduced in any form unless permitted by contract or by written permission of Cray Inc. U.S. GOVERNMENT RESTRICTED RIGHTS NOTICE The Computer Software is delivered as "Commercial Computer Software" as defined in DFARS 48 CFR 252.227-7014. All Computer Software and Computer Software Documentation acquired by or for the U.S. Government is provided with Restricted Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7014, as applicable. Technical Data acquired by or for the U.S. Government, if any, is provided with Limited Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7013, as applicable. Cray, LibSci, and PathScale are federally registered trademarks and Active Manager, Cray Apprentice2, Cray Apprentice2 Desktop, Cray C++ Compiling System, Cray CX, Cray CX1, Cray CX1-iWS, Cray CX1-LC, Cray CX1000, Cray CX1000-C, Cray CX1000-G, Cray CX1000-S, Cray CX1000-SC, Cray CX1000-SM, Cray CX1000-HN, Cray Fortran Compiler, Cray Linux Environment, Cray SHMEM, Cray X1, Cray X1E, Cray X2, Cray XD1, Cray XE, Cray XEm, Cray XE5, Cray XE5m, Cray XE6, Cray XE6m, Cray XK6, Cray XMT, Cray XR1, Cray XT, Cray XTm, Cray XT3, Cray XT4, Cray XT5, Cray XT5h, Cray XT5m, Cray XT6, Cray XT6m, CrayDoc, CrayPort, CRInform, ECOphlex, Gemini, Libsci, NodeKARE, RapidArray, SeaStar, SeaStar2, SeaStar2+, Sonexion, The Way to Better Science, Threadstorm, uRiKA, and UNICOS/lc are trademarks of Cray Inc.
    [Show full text]
  • Overview of Recent Supercomputers
    Overview of recent supercomputers Aad J. van der Steen high-performance Computing Group Utrecht University P.O. Box 80195 3508 TD Utrecht The Netherlands [email protected] www.phys.uu.nl/~steen NCF/Utrecht University July 2007 Abstract In this report we give an overview of high-performance computers which are currently available or will become available within a short time frame from vendors; no attempt is made to list all machines that are still in the development phase. The machines are described according to their macro-architectural class. Shared and distributed-memory SIMD an MIMD machines are discerned. The information about each machine is kept as compact as possible. Moreover, no attempt is made to quote price information as this is often even more elusive than the performance of a system. In addition, some general information about high-performance computer architectures and the various processors and communication networks employed in these systems is given in order to better appreciate the systems information given in this report. This document reflects the technical state of the supercomputer arena as accurately as possible. However, the author nor NCF take any responsibility for errors or mistakes in this document. We encourage anyone who has comments or remarks on the contents to inform us, so we can improve this report. NCF, the National Computing Facilities Foundation, supports and furthers the advancement of tech- nical and scientific research with and into advanced computing facilities and prepares for the Netherlands national supercomputing policy. Advanced computing facilities are multi-processor vectorcomputers, mas- sively parallel computing systems of various architectures and concepts and advanced networking facilities.
    [Show full text]
  • Massive Analytics for Streaming Graph Problems David A
    Massive Analytics for Streaming Graph Problems David A. Bader Outline •Overview • Cray XMT • Streaming Data Analysis – STINGER data structure • Tracking Clustering Coefficients • Tracking Connected Components • Parallel Graph Frameworks David A. Bader 2 STING Initiative: Focusing on Globally Significant Grand Challenges • Many globally-significant grand challenges can be modeled by Spatio- Temporal Interaction Networks and Graphs (or “STING”). • Emerging real-world graph problems include – detecting community structure in large social networks, – defending the nation against cyber-based attacks, – improving the resilience of the electric power grid, and – detecting and preventing disease in human populations. • Unlike traditional applications in computational science and engineering, solving these problems at scale often raises new research challenges because of sparsity and the lack of locality in the massive data, design of parallel algorithms for massive, streaming data analytics, and the need for new exascale supercomputers that are energy-efficient, resilient, and easy-to-program. David A. Bader 3 Center for Adaptive Supercomputing Software • WyomingClerk, launched July 2008 • Pacific-Northwest Lab – Georgia Tech, Sandia, WA State, Delaware • The newest breed of supercomputers have hardware set up not just for speed, but also to better tackle large networks of seemingly random data. And now, a multi-institutional group of researchers has been awarded over $14 million to develop software for these supercomputers. Applications include anywhere complex webs of information can be found: from internet security and power grid stability to complex biological networks. David A. Bader 4 CASS-MT TASK 7: Analysis of Massive Social Networks Objective To design software for the analysis of massive-scale spatio-temporal interaction networks using multithreaded architectures such as the Cray XMT.
    [Show full text]
  • Grappa: a Latency-Tolerant Runtime for Large-Scale Irregular Applications
    Grappa: A Latency-Tolerant Runtime for Large-Scale Irregular Applications Jacob Nelson, Brandon Holt, Brandon Myers, Preston Briggs, Luis Ceze, Simon Kahan, Mark Oskin University of Washington Department of Computer Science and Engineering fnelson, bholt, bdmyers, preston, luisceze, skahan, [email protected] Abstract works runs anywhere from a few to hundreds of microseconds – tens Grappa is a runtime system for commodity clusters of multicore of thousands of processor clock cycles. Since irregular application computers that presents a massively parallel, single address space tasks encounter remote references dynamically during execution and abstraction to applications. Grappa’s purpose is to enable scalable must resolve them before making further progress, stalls are frequent performance of irregular parallel applications, such as branch and and lead to severely underutilized compute resources. bound optimization, SPICE circuit simulation, and graph processing. While some irregular applications can be manually restructured Poor data locality, imbalanced parallel work and complex communi- to better exploit locality, aggregate requests to increase network cation patterns make scaling these applications difficult. message size, and manage the additional challenges of load balance Grappa serves both as a C++ user library and as a foundation for and synchronization, the effort required to do so is formidable and higher level languages. Grappa tolerates delays to remote memory by involves knowledge and skills pertaining to distributed systems far multiplexing thousands of lightweight workers to each processor core, beyond those of most application programmers. Luckily, many of balances load via fine-grained distributed work-stealing, increases the important irregular applications naturally offer large amounts communication throughput by aggregating smaller data requests of concurrency.
    [Show full text]
  • The Cray XMT™ Supercomputing System Is a Scalable Massively
    CRAY XMT PLATFORM OVERVIEW The Cray XMT™ supercomputing system is a scalable massively multithreaded platform with globally shared memory architecture for large-scale data analysis and data mining. The system is purpose-built for parallel applications that are dynamically changing, require random access to shared memory and typically do not run well on conventional systems. Multithreaded technology is ideally suited for tasks such as pattern matching, scenario development, behavioral prediction, anomaly identification and graph analysis. • Architected for Large-scale Data Analysis • Exploits the scalable Cray XT™ infrastructure • Scales from 24 to over 8000 processors providing over one million simultaneous threads and 128 terabytes of shared memory • UNICOS/lc operating system • Separately dedicated compute, service and I/O nodes • Incorporates custom Cray Threadstorm™ multithreaded processor using AMD Torrenza Open Socket technology Programming Environment The Cray XMT programming environment includes Architectural Overview advanced program analysis tools for software development and tuning: The Cray XMT system was architected to leverage Cray’s MPP system design to create a scalable, reliable and economical multithreaded supercomputing platform. The design • C and C++ compilers is based on a Cray MPP compute blade but utilizes AMD Torrenza Innovation Socket • Automatic parallelization technology to populate the AMD Opteron sockets with custom Cray Threadstorm chips developed for multithreaded processing. A single Cray Threadstorm processor can • STL and common math libraries sustain 128 simultaneous threads and is connected with up to 16 GB of memory that is globally accessible by any other processor in the system. • Cross compiler on Linux processors Each Cray Threadstorm processor is directly connected to a dedicated Cray SeaStar2™ • gcc/g++ on Linux nodes interconnect chip, resulting in a high bandwidth, low latency network characteristic of all • Cray Apprentice2™ for compiler analysis and Cray systems.
    [Show full text]
  • Exploring Datavortex Systems for Irregular Applications
    Exploring DataVortex Systems for Irregular Applications Roberto Gioiosa, Antonino Tumeo, Jian Yin, Thomas Warfel David Haglin Santiago Betelu Pacific Northwest National Laboratory Trovares Inc. Data Vortex Technologies Inc. froberto.gioiosa, antonino.tumeo, jian.yin, [email protected] [email protected] [email protected] Abstract—Emerging applications for data analytics and knowl- quickly execute regular computation patterns. Their network edge discovery typically have irregular or unpredictable com- interconnects are optimized for bulk-synchronous applications munication patterns that do not scale well on parallel systems characterized by large, batched data transfers and well-defined designed for traditional bulk-synchronous HPC applications. New network architectures that focus on minimizing (short) message communication patterns. Clusters optimized for traditional latencies, rather than maximizing (large) transfer bandwidths, HPC applications rarely demonstrate the same scalability for are emerging as possible alternatives to better support those irregular applications. applications with irregular communication patterns. We explore Several custom large-scale hardware designs (e.g., Cray a system based upon one such novel network architecture, XMT and Convey MX-100) [2], [3] have been proposed to the Data Vortex interconnection network, and examine how this system performs by running benchmark code written for better cope with the requirements of irregular applications but the Data Vortex network, as well as a reference MPI-over- are too expensive for general use. Software runtime layers Infiniband implementation, on the same cluster. Simple commu- (e.g., GMT, Grappa, Parallex, Active Pebbles) [4], [5], [6], nication primitives (ping-pong and barrier synchronization), a [7] have also looked into better support requirements of irreg- few common communication kernels (distributed 1D Fast Fourier ular applications by exploiting available hardware resources.
    [Show full text]
  • Cray Linux Environment™ (CLE) 4.0 Software Release Overview Supplement
    TM Cray Linux Environment™ (CLE) 4.0 Software Release Overview Supplement S–2497–4002 © 2010, 2011 Cray Inc. All Rights Reserved. This document or parts thereof may not be reproduced in any form unless permitted by contract or by written permission of Cray Inc. U.S. GOVERNMENT RESTRICTED RIGHTS NOTICE The Computer Software is delivered as "Commercial Computer Software" as defined in DFARS 48 CFR 252.227-7014. All Computer Software and Computer Software Documentation acquired by or for the U.S. Government is provided with Restricted Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7014, as applicable. Technical Data acquired by or for the U.S. Government, if any, is provided with Limited Rights. Use, duplication or disclosure by the U.S. Government is subject to the restrictions described in FAR 48 CFR 52.227-14 or DFARS 48 CFR 252.227-7013, as applicable. Cray, LibSci, and PathScale are federally registered trademarks and Active Manager, Cray Apprentice2, Cray Apprentice2 Desktop, Cray C++ Compiling System, Cray CX, Cray CX1, Cray CX1-iWS, Cray CX1-LC, Cray CX1000, Cray CX1000-C, Cray CX1000-G, Cray CX1000-S, Cray CX1000-SC, Cray CX1000-SM, Cray CX1000-HN, Cray Fortran Compiler, Cray Linux Environment, Cray SHMEM, Cray X1, Cray X1E, Cray X2, Cray XD1, Cray XE, Cray XEm, Cray XE5, Cray XE5m, Cray XE6, Cray XE6m, Cray XK6, Cray XMT, Cray XR1, Cray XT, Cray XTm, Cray XT3, Cray XT4, Cray XT5, Cray XT5h, Cray XT5m, Cray XT6, Cray XT6m, CrayDoc, CrayPort, CRInform, ECOphlex, Gemini, Libsci, NodeKARE, RapidArray, SeaStar, SeaStar2, SeaStar2+, Sonexion, The Way to Better Science, Threadstorm, uRiKA, and UNICOS/lc are trademarks of Cray Inc.
    [Show full text]
  • Research Landscape
    RESEARCH LANDSCAPE Ohio’s strengths in basic and applied research are broad and deep, spanning a multitude of academic, business and industrial interests. The spectrum of clients served by the Ohio Supercomputer Center encompasses many fields of study, from establishing secure, firewall-free network architecture to improving injection molds for manufacturing, as highlighted on the following pages. PARALLEL COMPARISONS Bokhari conducts comparisons of system architectures Surveying the wide range of parallel system architectures offered in the supercomputer market, a former Ohio State University graduate research student sought to establish some side-by-side performance comparisons. “We explored the parallelization of the subset-sum problem by the sum of their sizes. Also, she carefully timed the code on three contemporary but very different architectures: a runs for solving a comprehensive range of problem sizes. 128-processor Cray massively multithreaded machine, a Bokhari’s results illustrate that the subset-sum problem can 16-processor IBM shared memory machine and a 240-core be parallelized well on all three architectures, although for NVIDIA graphics processing unit,” said former computer different ranges of problem sizes. Bokhari concluded that the science and engineering graduate student Saniyah Bokhari. graphical processing unit (GPU) architecture performs well “These experiments highlighted the strengths and weaknesses for problems whose tables fit within the limitations of the of these architectures in the context of a well-defined combi- device memory. Because GPUs typically have memory sizes natorial problem.” in the range of 10 gigabytes (GB), such architectures are best Bokhari evaluated the conventional central processing unit for small problems that have table sizes of approximately 30 architecture of the IBM 1350 Glenn Cluster at the Ohio billion bits.
    [Show full text]