Online Modeling and Tuning of Parallel Stream Processing Systems Jonathan Curtis Beard Washington University in St

Total Page:16

File Type:pdf, Size:1020Kb

Online Modeling and Tuning of Parallel Stream Processing Systems Jonathan Curtis Beard Washington University in St Washington University in St. Louis Washington University Open Scholarship Engineering and Applied Science Theses & McKelvey School of Engineering Dissertations Summer 8-15-2015 Online Modeling and Tuning of Parallel Stream Processing Systems Jonathan Curtis Beard Washington University in St. Louis Follow this and additional works at: https://openscholarship.wustl.edu/eng_etds Part of the Engineering Commons Recommended Citation Beard, Jonathan Curtis, "Online Modeling and Tuning of Parallel Stream Processing Systems" (2015). Engineering and Applied Science Theses & Dissertations. 125. https://openscholarship.wustl.edu/eng_etds/125 This Dissertation is brought to you for free and open access by the McKelvey School of Engineering at Washington University Open Scholarship. It has been accepted for inclusion in Engineering and Applied Science Theses & Dissertations by an authorized administrator of Washington University Open Scholarship. For more information, please contact [email protected]. WASHINGTON UNIVERSITY IN ST. LOUIS School of Engineering and Applied Science Department of Computer Science and Engineering Dissertation Examination Committee: Roger Dean Chamberlain, Chair Jeremy Buhler Ron Cytron Roch Guerin Jenine Harris Juan Pantano Robert B. Pless Online Modeling and Tuning of Parallel Stream Processing Systems by Jonathan Curtis Beard A dissertation presented to the Graduate School of Arts and Sciences of Washington University in partial fulfillment of the requirements for the degree of Doctor of Philosophy August 2015 Saint Louis, Missouri c 2015, Jonathan Curtis Beard Table of Contents ListofFigures.................................... v ListofTables .................................... viii Acknowledgements ................................. ix Abstract ....................................... xi Chapter1: Introduction.............................. 1 1.1 IndustryTrends.................................. 2 1.2 ParadigmFlux .................................. 3 1.3 TurningStreamsintoTorrents . .... 5 1.4 ContributionandStructure . ... 8 Chapter2: BackgroundandRelatedWorks . 10 2.1 StreamProcessing ................................ 10 2.2 Modeling...................................... 12 2.3 Instrumentation................................. 14 2.4 OnlineModeling&PerformanceTuning . .... 15 2.5 DynamicAdaptation ............................... 16 Chapter3: RaftLibStreamingLibrary . 18 3.1 Designconsiderations. ... 20 3.2 RaftLibdescription .............................. .. 21 3.2.1 RaftLibasaresearchplatform. .. 24 3.2.2 Authoringstreamingapplications . ... 29 3.3 Benchmarking................................... 33 3.4 ConcludingRemarksandWhatFollows. ... 37 Chapter 4: Modeling Streaming Applications . 38 4.1 Introduction.................................... 38 4.1.1 Description ................................ 39 4.1.2 SharingModels .............................. 45 4.1.3 ModelingAssumptions . 45 4.1.4 Example.................................. 46 ii 4.2 ModelEvaluationApproach . 48 4.2.1 Tools.................................... 48 4.2.2 Hardware ................................. 48 4.2.3 EmpiricalTesting............................. 49 4.2.4 Selecting Compute Resources and Mapping Application Kernels . 49 4.2.5 SyntheticBenchmarks . .. 50 4.2.6 RealApplications............................. 51 4.3 Results....................................... 51 4.3.1 ProcessorSharingModel . 51 4.3.2 TheFlowModel ............................. 52 4.3.3 TheQueueingModel ........................... 54 4.4 Conclusions .................................... 56 Chapter5: BestCaseExecutionTimeVariation . 57 5.1 Methodology ................................... 60 5.1.1 SyntheticWorkload............................ 60 5.1.2 Hardware, Software, and Data Collection . .... 61 5.1.3 Distribution................................ 64 5.1.4 Parameterization and definition of mLevy ............... 67 5.2 Results....................................... 69 5.3 Conclusions .................................... 72 Chapter6: DynamicInstrumentation. 75 6.1 Introduction.................................... 75 6.2 InstrumentationConsiderations . ...... 79 6.2.1 Throughput ................................ 82 6.2.2 QueueOccupancy............................. 82 6.3 ServiceRate.................................... 83 6.3.1 OnlineServiceRateHeuristic . 85 6.3.2 SamplingPeriodDetermination . .. 85 6.3.3 ServiceRateHeuristic .......................... 86 6.4 Evaluation..................................... 92 6.4.1 Infrastructure ............................... 92 6.5 Applications.................................... 92 6.5.1 MatrixMultiply.............................. 93 6.5.2 Rabin-KarpStringSearch . 94 6.6 Validation ..................................... 94 6.7 Conclusions .................................... 99 Chapter7: ModelSelection............................ 101 7.1 Stochastic Models and Streaming Applications . ........ 103 7.2 StochasticQueueingModelSelection . ...... 104 7.2.1 Methodology ............................... 104 iii 7.2.2 SupportVectorMachine . 105 7.2.3 DataCollection&Hardware. 107 7.2.4 SVMandTraining ............................ 109 7.2.5 Artificial Neural Network (ANN) . 113 7.3 Conclusions&FutureWork . 114 Chapter 8: Online Tuning: Putting It All Together . 117 8.1 OnlineModelingofStreamingSystems . ..... 117 8.1.1 Whyisonlinetuningimportant? . 119 8.1.2 AdaptionTypes.............................. 119 8.1.3 EvaluationMethodology . 121 8.1.4 AdaptionResults ............................. 121 8.2 Conclusions .................................... 124 Chapter 9: Conclusions and Future Work . 125 9.1 Conclusions .................................... 125 9.2 FutureWork.................................... 127 References ...................................... 130 Vita.......................................... 146 iv List of Figures Figure3.1: Simplestreamingexample . .... 19 Figure3.2: RaftLibsumkernelexample. .... 23 Figure 3.3: RaftLib sum application mapping example . ........ 24 Figure3.4: RaftLibscheduleroptions . ..... 26 Figure3.5: Importance ofproperly sized buffers . ....... 27 Figure 3.6: RaftLib syntax for C++ container interaction . .......... 31 Figure3.7: RaftLibforeachconstruct . ..... 32 Figure3.8: RaftLiblambdakernelsyntax . ..... 33 Figure3.9: RaftLibstringmatchingtopology . ....... 34 Figure 3.10: Aho-Corasick string matching algorithm . ......... 34 Figure 3.11: RaftLib performance by thread count. ........ 36 Figure4.1: JPEGencodetopology. 40 Figure4.2: Stagesofflowmodeltransformation . ...... 40 Figure4.3: Exampleapplicationtopology . ...... 46 Figure4.4: Applicationmapping. 46 Figure 4.5: Traditional service rate characterization . ............ 47 Figure4.6: Flowmodelingstepbystep . .. 47 Figure4.7: Completedsolutionstoflowmodel . ..... 47 v Figure 4.8: Offline service rate characterization . ......... 49 Figure4.9: DESapplicationtopology . .... 51 Figure4.10: Percenterrorforsharingmodel . ....... 52 Figure 4.11: Synthetic app % error for gain/loss flow model . ......... 53 Figure 4.12: JPEG encode % error for gain/loss flow model. ....... 53 Figure 4.13: DES encrypt % error for gain/loss flow model . ........ 53 Figure 4.14: Synthetic app buffer capacity % error . ....... 54 Figure4.15: JPEGencodebuffercapacity%error. ..... 55 Figure4.16: DESencryptbuffer capacity %error . ...... 55 Figure5.1: Histogramofexecutionvariation . ...... 58 Figure 5.2: Simplifying distributional assumptions . ........... 59 Figure5.3: Memoryaccessvariances. .... 63 Figure5.4: Timervariance ............................. 64 Figure5.5: Timerlatency.............................. 65 Figure 5.6: mLevyQQ-plotsgroupedbyprocess . 71 Figure5.7: KLdivergenceformodelselection. ....... 72 Figure 6.1: Non-blocking service rate intuition . ........ 77 Figure 6.2: Non-blocked periods in highly utilized queue . .......... 78 Figure6.3: Twoserverstreamingsystem . .... 79 Figure 6.4: Instrumentation monitor arrangement . ........ 80 Figure6.5: TimingacrossNUMAnodes. 81 Figure6.6: Stabilityoftimeframe . .... 82 Figure 6.7: Probability of observing non-blocking read . ........... 84 vi Figure 6.8: Direct observations of service rate. ......... 87 Figure 6.9: Initial filtered observations of non-blocked servicerate ........ 89 Figure 6.10: Convergence of stable non-blocking service rates............ 90 Figure 6.11: Estimating the point of convergence . ........ 91 Figure 6.12: Instrumentation of shifting distributions . ............. 91 Figure6.13: Matrixmultiplyexample. ..... 93 Figure6.14: Rabin-Karpexample . ... 94 Figure 6.15: Histogram of micro-benchmark instrumentation . ........... 95 Figure 6.16: Micro-benchmark shifting distribution results ............. 96 Figure6.17: Detectionofbothphases. ..... 97 Figure 6.18: Matrix multiply instrumentation . ........ 98 Figure 6.19: Rabin-Karp instrumentation results . ......... 99 Figure 7.1: Fingerprinting of computer systems . ....... 102 Figure7.2: Simplemicro-benchmark. .... 103 Figure7.3: Machinelearningfeaturecloud . ...... 106 Figure7.4: Classificationbyerrorcategory . ....... 110 Figure 7.5: Classification rate by queue utilization (ρ)............... 111 Figure7.6: Neuralnetworkresults . 115 Figure 8.1: Complete enumeration of RaftLib application . .......... 122 Figure8.2: Peakperformancezoom-in. .... 123 Figure 8.3: Parallelization event monitor control actions . ............. 123 Figure 8.4: Queue sizing event monitor control actions . .......... 123 vii List of Tables Table 3.1: Benchmarking hardware summary for RaftLib . .......
Recommended publications
  • CRN What It Was Doing and Why It Was Cognitive Systems Vision Doing It, and to Recover from Mental Continued on Page 8 Expanding the Pipeline
    COMPUTING RESEARCH NEWS Computing Research Association, Celebrating 30 Years of Service to the Computing Research Community November 2002 Vol. 14/No. 5 DARPA’s New Cognitive Systems Vision By Ron Brachman and IPTO’s goal is to create a new to cope with systems both keep Zachary Lemnios generation of cognitive systems. growing. In order to make our systems more reliable, more secure, The impact of the Defense Mired in Moore’s Law? and more understandable, and to Advanced Research Projects Agency One benefit of such cognitive continue making substantial contri- (DARPA) on computing over the systems would be their help in butions to society, we need to do past 40 years has been profound. Led extracting us from a corner into something dramatically different. by the visionary J.C.R. Licklider and which our success seems to have his innovative successors in the painted us. The research that has The Promise of Cognitive Information Processing Techniques helped the industry follow Moore’s Systems Office (IPTO), DARPA initiated “Law” has created processors that are IPTO is attacking this problem by work that ultimately put personal remarkably fast and small, and data driving a fundamental change in computers on millions of desktops storage capabilities that are vast and computing systems. By giving systems Ron Brachman and made the global Internet a cheap. Unfortunately, these incred- more cognitive capabilities, we reality. In fact, the original IPTO, ible developments have cut two ways. believe we can make them more which lasted from 1962 to 1985, was While today’s computers are more responsible for their own behavior in large part responsible for estab- powerful than ever, we have been and maintenance.
    [Show full text]
  • The Conference Program Booklet
    Austin Convention Center Conference Austin, TX Program http://sc15.supercomputing.org/ Conference Dates: Exhibition Dates: The International Conference for High Performance Nov. 15 - 20, 2015 Nov. 16 - 19, 2015 Computing, Networking, Storage and Analysis Sponsors: SC15.supercomputing.org SC15 • Austin, Texas The International Conference for High Performance Computing, Networking, Storage and Analysis Sponsors: 3 Table of Contents Welcome from the Chair ................................. 4 Papers ............................................................... 68 General Information ........................................ 5 Posters Research Posters……………………………………..88 Registration and Conference Store Hours ....... 5 ACM Student Research Competition ........ 114 Exhibit Hall Hours ............................................. 5 Posters SC15 Information Booth/Hours ....................... 5 Scientific Visualization/ .................................... 120 Data Analytics Showcase SC16 Preview Booth/Hours ............................. 5 Student Programs Social Events ..................................................... 5 Experiencing HPC for Undergraduates ...... 122 Registration Pass Access .................................. 7 Mentor-Protégé Program .......................... 123 Student Cluster Competition Kickoff ......... 123 SCinet ............................................................... 8 Student-Postdoc Job & ............................. 123 Convention Center Maps ................................. 12 Opportunity Fair Daily Schedules
    [Show full text]
  • High-Level and Efficient Stream Parallelism on Multi-Core Systems
    WSCAD 2017 - XVIII Simp´osio em Sistemas Computacionais de Alto Desempenho High-Level and Efficient Stream Parallelism on Multi-core Systems with SPar for Data Compression Applications Dalvan Griebler1, Renato B. Hoffmann1, Junior Loff1, Marco Danelutto2, Luiz Gustavo Fernandes1 1 Faculty of Informatics (FACIN), Pontifical Catholic University of Rio Grande do Sul (PUCRS), GMAP Research Group, Porto Alegre, Brazil. 2Department of Computer Science, University of Pisa (UNIPI), Pisa, Italy. {dalvan.griebler, renato.hoffmann, junior.loff}@acad.pucrs.br, [email protected], [email protected] Abstract. The stream processing domain is present in several real-world appli- cations that are running on multi-core systems. In this paper, we focus on data compression applications that are an important sub-set of this domain. Our main goal is to assess the programmability and efficiency of domain-specific language called SPar. It was specially designed for expressing stream paral- lelism and it promises higher-level parallelism abstractions without significant performance losses. Therefore, we parallelized Lzip and Bzip2 compressors with SPar and compared with state-of-the-art frameworks. The results revealed that SPar is able to efficiently exploit stream parallelism as well as provide suit- able abstractions with less code intrusion and code re-factoring. 1. Introduction Over the past decade, vendors realized that increasing clock frequency to gain perfor- mance was no longer possible. Companies were then forced to slow the clock frequency and start adding multiple processors to their chips. Since that, software started to rely on parallel programming to increase performance [Sutter 2005]. However, exploiting par- allelism in such multi-core architectures is a challenging task that is still too low-level and complex for application programmers.
    [Show full text]
  • Curriculum Vitae
    Massachusetts Institute of Technology School of Engineering Faculty Personnel Record Date: April 1, 2020 Full Name: Charles E. Leiserson Department: Electrical Engineering and Computer Science 1. Date of Birth November 10, 1953 2. Citizenship U.S.A. 3. Education School Degree Date Yale University B. S. (cum laude) May 1975 Carnegie-Mellon University Ph.D. Dec. 1981 4. Title of Thesis for Most Advanced Degree Area-Efficient VLSI Computation 5. Principal Fields of Interest Analysis of algorithms Caching Compilers and runtime systems Computer chess Computer-aided design Computer network architecture Digital hardware and computing machinery Distance education and interaction Fast artificial intelligence Leadership skills for engineering and science faculty Multicore computing Parallel algorithms, architectures, and languages Parallel and distributed computing Performance engineering Scalable computing systems Software performance engineering Supercomputing Theoretical computer science MIT School of Engineering Faculty Personnel Record — Charles E. Leiserson 2 6. Non-MIT Experience Position Date Founder, Chairman of the Board, and Chief Technology Officer, Cilk Arts, 2006 – 2009 Burlington, Massachusetts Director of System Architecture, Akamai Technologies, Cambridge, 1999 – 2001 Massachusetts Shaw Visiting Professor, National University of Singapore, Republic of 1995 – 1996 Singapore Network Architect for Connection Machine Model CM-5 Supercomputer, 1989 – 1990 Thinking Machines Programmer, Computervision Corporation, Bedford, Massachusetts 1975
    [Show full text]
  • Queue Streaming Model Theory, Algorithms, and Implementation
    Mississippi State University Scholars Junction Theses and Dissertations Theses and Dissertations 1-1-2019 Queue Streaming Model Theory, Algorithms, and Implementation Anup D. Zope Follow this and additional works at: https://scholarsjunction.msstate.edu/td Recommended Citation Zope, Anup D., "Queue Streaming Model Theory, Algorithms, and Implementation" (2019). Theses and Dissertations. 3708. https://scholarsjunction.msstate.edu/td/3708 This Dissertation - Open Access is brought to you for free and open access by the Theses and Dissertations at Scholars Junction. It has been accepted for inclusion in Theses and Dissertations by an authorized administrator of Scholars Junction. For more information, please contact [email protected]. Queue streaming model: Theory, algorithms, and implementation By Anup D. Zope A Dissertation Submitted to the Faculty of Mississippi State University in Partial Fulfillment of the Requirements for the Degree of Doctor of Philosophy in Computational Engineering in the Center for Advanced Vehicular Systems Mississippi State, Mississippi May 2019 Copyright by Anup D. Zope 2019 Queue streaming model: Theory, algorithms, and implementation By Anup D. Zope Approved: Edward A. Luke (Major Professor) Shanti Bhushan (Committee Member) Maxwell Young (Committee Member) J. Mark Janus (Committee Member) Seongjai Kim (Committee Member) Manav Bhatia (Graduate Coordinator) Jason M. Keith Dean Bagley College of Engineering Name: Anup D. Zope Date of Degree: May 3, 2019 Institution: Mississippi State University Major Field: Computational Engineering Major Professor: Edward A. Luke Title of Study: Queue streaming model: Theory, algorithms, and implementation Pages of Study: 195 Candidate for Degree of Doctor of Philosophy In this work, a model of computation for shared memory parallelism is presented.
    [Show full text]
  • Pash: Light-Touch Data-Parallel Shell Processing
    PaSh: Light-touch Data-Parallel Shell Processing Nikos Vasilakis∗ Konstantinos Kallas∗ Konstantinos Mamouras MIT University of Pennsylvania Rice University [email protected] [email protected] [email protected] Achilles Benetopoulos Lazar Cvetković Unaffiliated University of Belgrade [email protected] [email protected] Abstract Parallelizability Parallelizing Runtime Classes §3 Transformations §4.3 Primitives §5 Dataflow This paper presents PaSh, a system for parallelizing POSIX POSIX, GNU §3.1 Regions shell scripts. Given a script, PaSh converts it to a dataflow Annotations §3.2 § 4.1 DFG § 4.4 graph, performs a series of semantics-preserving program §4.2 transformations that expose parallelism, and then converts Seq. Script Par. Script the dataflow graph back into a script—one that adds POSIX constructs to explicitly guide parallelism coupled with PaSh- Fig. 1. PaSh overview. PaSh identifies dataflow regions (§4.1), converts provided Unix-aware runtime primitives for addressing per- them to dataflow graphs (§4.2), applies transformations (§4.3) based onthe parallelizability properties of the commands in these regions (§3.1, §3.2), formance- and correctness-related issues. A lightweight an- and emits a parallel script (§4.4) that uses custom primitives (§5). notation language allows command developers to express key parallelizability properties about their commands. An accompanying parallelizability study of POSIX and GNU • Command developers, responsible for implementing indi- commands—two large and commonly used groups—guides vidual commands such as sort, uniq, and jq. These de- the annotation language and optimized aggregator library velopers usually work in a single programming language, that PaSh uses. PaSh’s extensive evaluation over 44 unmod- leveraging its abstractions to provide parallelism when- ified Unix scripts shows significant speedups (0.89–61.1×, ever possible.
    [Show full text]
  • A Personal Analytics Platform for the Internet of Things Implementing Kappa Architecture with Microservice-Based Stream Processing
    A Personal Analytics Platform for the Internet of Things Implementing Kappa Architecture with Microservice-based Stream Processing Theo Zschörnig1, Robert Wehlitz1 and Bogdan Franczyk2,3 1Institute for Applied Informatics (InfAI), Leipzig University, Hainstr. 11, 04109 Leipzig, Germany 2Information Systems Institute, Leipzig University, Grimmaische Str. 12, 04109 Leipzig, Germany 3Business Informatics Institute, Wrocław University of Economics, ul. Komandorska 118-120, 53-345 Wrocław, Poland Keywords: Personal Analytics, Internet of Things, Kappa Architecture, Microservices, Stream Processing. Abstract: The foundation of the Internet of Things (IoT) consists of different devices, equipped with sensors, actuators and tags. With the emergence of IoT devices and home automation, advantages from data analysis are not limited to businesses and industry anymore. Personal analytics focus on the use of data created by individuals and used by them. Current IoT analytics architectures are not designed to respond to the needs of personal analytics. In this paper, we propose a lightweight flexible analytics architecture based on the concept of the Kappa Architecture and microservices. It aims to provide an analytics platform for huge numbers of different scenarios with limited data volume and different rates in data velocity. Furthermore, the motivation for and challenges of personal analytics in the IoT are laid out and explained as well as the technological approaches we use to overcome the shortcomings of current IoT analytics architectures. 1 INTRODUCTION also the opportunities they provide (Section 2). We give an overview of the state of the art in IoT analytics It is estimated that the number of Internet of Things regarding technologies and architectures and show (IoT) devices will grow in huge quantities, to around that these are not fully suitable for personal analytics 24 billion in the year 2020 (Greenough, 2016).
    [Show full text]
  • Computational Thinking
    0 Computational Thinking Jeannette M. Wing President’s Professor of Computer Science and Department Head Computer Science Department Carnegie Mellon University Microsoft Asia Faculty Summit 26 October 2012 Tianjin, China My Grand Vision • Computational thinking will be a fundamental skill used by everyone in the world by the middle of the 21st Century. – Just like reading, writing, and arithmetic. – Incestuous: Computing and computers will enable the spread of computational thinking. – In research: scientists, engineers, …, historians, artists – In education: K-12 students and teachers, undergrads, … J.M. Wing, “Computational Thinking,” CACM Viewpoint, March 2006, pp. 33-35. Paper off http://www.cs.cmu.edu/~wing/ Computational Thinking 2 Jeannette M. Wing Computing is the Automation of Abstractions Abstractions 1. Machine 2. Human Automation 3. Network [Machine + Human] Computational Thinking focuses on the process of abstraction - choosing the right abstractions - operating in terms of multiple layers of abstraction simultaneously as in - defining the relationships the between layers Mathematics guided by the following concerns… Computational Thinking 3 Jeannette M. Wing Measures of a “Good” Abstraction in C.T. as in • Efficiency Engineering – How fast? – How much space? NEW – How much power? • Correctness – Does it do the right thing? • Does the program compute the right answer? – Does it do anything? • Does the program eventually produce an answer? [Halting Problem] • -ilities – Simplicity and elegance – Scalability – Usability – Modifiability – Maintainability – Cost – … Computational Thinking 4 Jeannette M. Wing Computational Thinking, Philosophically • Complements and combines mathematical and engineering thinking – C.T. draws on math as its foundations • But we are constrained by the physics of the underlying machine – C.T.
    [Show full text]
  • Research News
    Computing Research News COMPUTING RESEARCH ASSOCIATION, CELEBRATING 40 YEARS OF SERVICE TO THE COMPUTING RESEARCH COMMUNITY APRIL 2014 Vol. 26 / No. 4 Announcements 2 2014 CRA Board Election Results 3 Visions 2025: Interacting with the Computers All Around Us 4 Conference at Snowbird 4 CERP Infographic 5 Postdoc Best Practices Award Recipients Announced 6 University-Industry Partnership to Advance Machine Learning 7 Expanding the Pipeline: 1st CRA-W/CDC Broadening Participation in Visualization (BPViz) Workshop 8 CCC Workshop Report: Multidisciplinary Research for Online Education 10 Highlights of the CISE Fiscal Year 2015 Budget Request 11 Time to Degree in Computing 13 CRA Board Members 16 CRA Board Officers 16 CRA Staff 16 Professional Opportunities 17 COMPUTING RESEARCH NEWS, APRIL 2014 Vol. 26 / No. 4 Announcements Rabin and Klawe Named 2014 Women of Vision by the Richard Tapia receives Anita Borg Institute 2014 Vannevar Bush Award CRA, CRA-W and CDC congratulate Richard Tapia for receiving the 2014 Vannevar Bush Award. National Science Board has announced that mathematician Photo credit – Rice University Richard Tapia, a leader in mentoring minorities in science, engineering Dr. Tal Rabin Dr. Maria Klawe Richard Tapia and mathematics fields, is the 2014 recipient Congratulations to both Tal Rabin and Maria Klawe. Klawe is of its Vannevar Bush Award. Tapia is also a previous president of Harvey Mudd College and was a founding co- recipient of CRA’s A. Nico Habermann Award, and the chair of the highly successful CRA-W Committee. She will be Richard Tapia Celebration of Diversity in Computing a plenary speaker at the 2014 Conference at Snowbird.
    [Show full text]
  • Paving the Way Towards High-Level Parallel Pattern Interfaces for Data Stream Processing
    This is a postprint version of the following published document: Del Río Astorga, D., Dolz, M. F., Fernández, J., García, J.D., (2018). Paving the way towards high- level parallel pattern interfaces for data stream processing. Future Generation Computer Systems, 87, pp. 228-241. DOI: https://doi.org/10.1016/j.future.2018.05.011 © 2018 Elsevier B.V. All rights reserved. This work is licensed under a Creative Commons Attribution-NonCommercial- NoDerivatives 4.0 International License. Paving the Way Towards High-Level Parallel Pattern Interfaces for Data Stream Processing David del Rio Astorga1, Manuel F. Dolz1, Javier Fernandez´ 1 and J. Daniel Garc´ıa1 1 Universidad Carlos III de Madrid, 28911–Leganes,´ Spain The emergence of data stream applications has posed a number of new challenges to existing infrastruc- tures, processing engines and programming models. In this sense, high-level interfaces, encapsulating algorithmic aspects in pattern-based constructions, have considerable reduced the development and par- allelization efforts of these type of applications. An example of parallel pattern interface is GrPPI, a C++ generic high-level library that acts as a layer between developers and existing parallel programming frameworks, such as C++ threads, OpenMP and Intel TBB. In this paper, we complement the basic pat- terns supported by GrPPI with the new stream operators Split-Join and Window, and the advanced parallel patterns Stream-Pool, Windowed-Farm and Stream-Iterator for the aforementioned back ends. Thanks to these new stream operators, complex compositions among streaming patterns can be expressed. On the other hand, the collection of advanced patterns allows users to tackle some domain-specific applications, ranging from the evolutionary to the real-time computing areas, where compositions of basic patterns are not capable of fully mimicking the algorithmic behavior of their original sequential codes.
    [Show full text]
  • Parallel Discrete Event Simulation: the Making of a Field
    Proceedings of the 2017 Winter Simulation Conference W. K. V. Chan, A. D'Ambrogio, G. Zacharewicz, N. Mustafee, G. Wainer, and E. Page, eds. PARALLEL DISCRETE EVENT SIMULATION: THE MAKING OF A FIELD Richard M. Fujimoto Rajive Bagrodia School of Computational Science and Engineering Scalable Network Technologies Inc. Georgia Institute of Technology 6059 Bristol Parkway #200, Culver City, CA 90230 266 Ferst Drive Computer Science Dept., University of California Atlanta, GA 30332, USA Los Angeles, CA 90095, USA Randal E. Bryant K. Mani Chandy Computer Science Department Computing and Mathematical Sciences Carnegie Mellon University Department 5000 Forbes Ave California Institute of Technology Pittsburgh, PA 15213, USA Pasadena, CA 91125, USA David Jefferson Jayadev Misra Lawrence Livermore National Laboratory Computer Science Department 7000 East Avenue The University of Texas at Austin Livermore, CA 94550, USA Austin, TX 78712-1757, USA David Nicol Brian Unger Department of Electrical and Computer Science Department Computer Engineering University of Calgary University of Illinois, Urbana Champaign 2500 University Drive 1308 W. Main St., Urbana IL 61820, USA Calgary, Alberta T2N 1N4 Canada ABSTRACT Originating in the 1970’s, the parallel discrete event simulation (PDES) field grew from a group of researchers focused on determining how to execute a discrete event simulation program on a parallel computer while still obtaining the same results as a sequential execution. Over the decades that followed the field expanded, grew, and flourishes to this day. This paper describes the origins and development of the field in the words of many who were deeply involved. Unlike other published work focusing on technical issues, the emphasis here is on historical aspects that are not recorded elsewhere, providing a unique characterization of how the field was created and developed.
    [Show full text]
  • A Complete Bibliography of Publications in the International
    A Complete Bibliography of Publications in The International Journal of Supercomputer Applications, The International Journal of Supercomputer Applications and High-Performance Computing, and The International Journal of High Performance Computing Applications Nelson H. F. Beebe University of Utah Department of Mathematics, 110 LCB 155 S 1400 E RM 233 Salt Lake City, UT 84112-0090 USA Tel: +1 801 581 5254 FAX: +1 801 581 4148 E-mail: [email protected], [email protected], [email protected] (Internet) WWW URL: http://www.math.utah.edu/~beebe/ 18 May 2021 Version 2.03 Title word cross-reference IGDQO19, MC21, YZC+15]. M × N [BYCB05, DP05, JLO05]. N [HJ96, SWW94, DAB+12, FT19, RTRG+07, INS+20]. ? 2 [VSS+13, Wal18]. 2564 [IKY+10]. 3 [NC18]. [AKW19, ARR99, BGM15, CSGM17, + + FIMU19, GGS01, HSLK11, KC18, PR95, -Body [HJ96, SWW94, INS 20, RTRG 07, + SSCF19, THL88]. 2 [AMC+18]. 3 [KPST18]. DAB 12, FT19]. -D α [TKSK88]. d = 2 [BRT+92]. H [YIYD19]. [SSCF19, ARR99, GGS01, PR95, THL88]. * ∗ * -matrix [YIYD19]. CH + H2 ) CH3 ) CH2 + H [ASW91]. CuO2 [SSSW91]. ILU [SZ11]. k [TNLP13]. 0th K2 [CBW95]. LU [BLRR01, DD89, DD91, [RAGW93]. 1 2 100 [IHMM87]. 100k [INY+14]. 10P Accessing [HLP+03]. Accuracy [DD89]. 164 [LT88]. 1917-1991 [Mar91]. [HCCG20, SK20]. Accurate 1A [HXW+13]. 1s [McN89]. [LR09, LRO10, TMWS91, VFD04, WSD+14]. Acknowledgments [Ano02c, Ano02d]. 2 [CL95, CC95, CLM+16, DD89, DD91, Acoustic [GKN+96, MBF+11, ALE+20]. GCL93, HGMW12, LT90, LYL+16, Mor89a]. Across [KS05, SB19]. ACS [Mar88a]. 200/VF [DD89]. 2000 [CLBS17]. 2003 Action [DBA+09, Num04]. Active [Her91]. [OL05].
    [Show full text]