Shared-Variable Synchronization Approaches for Dynamic Dataflow

Total Page:16

File Type:pdf, Size:1020Kb

Shared-Variable Synchronization Approaches for Dynamic Dataflow Shared-variable synchronization approaches for dynamic dataflow programs Apostolos Modas1, Simone Casale-Brunet2, Robert Stewart3, Endri Bezati2, Junaid Ahmad4∗, Marco Mattavelli1 1EPFL SCI STI MM, Ecole´ Polytechnique Fed´ erale´ de Lausanne, Switzerland 2SIB Swiss Institute of Bioinformatics, Lausanne, Switzerland 3Mathematical and Computer Sciences, Heriot-Watt University, Edinburgh, United Kingdom 4J Nomics, Manchester, United Kingdom Abstract—This paper presents shared-variable synchronization architectural components with low-level, target specific code approaches for dataflow programming. The mechanisms do not synthesis. require any substantial model of computation (MoC) modifi- By relying on isolation and non-sharing, an actor can access cation, and is portable across both for hardware (HW) and software (SW) low-level code synthesis. With the shared-variable its own state without fear of data-races. The approach can formalization, the benefits of the dataflow MoC are maintained, introduce inefficiencies in generated code, e.g. allocations however the space and energy efficiency of an application can be causing high memory use, and high levels of actor idle time significantly improved. The approach targets Dynamic Process whilst data is explicitly copied. As an example, in the context Network (DPN) dataflow applications, thus making them also of video compression, the output data generally depends on suitable for less expressive models e.g. synchronous and cyclo- static dataflow that DPN subsumes. The approach is validated intermediate data structures that different actors are obliged to through the analysis and optimization of a High-Efficiency Video replicate since they cannot be shared. This can severely impact Coding (HEVC) decoder implemented in the RVC-CAL dataflow both the time and space performance and the energy efficiency language targeting a multi-core platform. Experimental results of an application. show how, starting from an initial design that does not use In this paper, we present a set of shared-variable synchro- the shared-variable formalism, frames per second throughput performance is increased by a factor of 21. nization approaches that do not require any substantial MoC modification and that are portable both for HW and SW low- I. INTRODUCTION level code synthesis. The main advantage of this formalization In recent years, there has been a renewed interest in the is the fact that the benefits of the dataflow MoC are main- field of dataflow programming. This has been driven by the tained, however, the space and time performance and energy limitation of the frequency increases of deep sub-micron efficiency of an application can be significantly improved. CMOS silicon technology, which has shifted the evolution of The method targets Dynamic Process Network (DPN) dataflow processing platforms to systems comprising heterogeneous ar- applications, thus making them also suitable for less expressive rays of parallel processors. The emergence of these manycore models e.g. synchronous and cyclo-static dataflow that DPN architectures poses new problems and challenges for compiling subsumes. applications efficiently to them. A major challenge is the The paper is structured as follows: Section II provides portability of applications across heterogeneous architectures, an overview of current dataflow MoCs and shared-variable in particular, the portability of the parallelism present in approaches. Section III presents a novel and generic shared each specific application [1]. Dataflow programming, in all variable paradigm that can be introduced in a dataflow MoC its different model of computations (MoCs), is well placed to without fear of data-races. Section IV describes how this overcome the challenges of exploiting efficient heterogeneous methodology has been implemented in the standardized RVC- architectures. Dataflow MoCs are widely used for the specifi- CAL dataflow language. Experimental results are provided in cation of data-driven algorithms in many application areas, e.g. Section V, where an HEVC video decoder implemented using video and audio processing, bioinformatics, financial trading, the RVC-CAL dataflow language has been optimized with the packet switching applications. In these domains, scalability use of shared variables. Finally, Section VI concludes the paper and composability of systems are increasingly important re- and discusses future work directions. quirements, and dataflow MoCs are an efficient way of imple- II. BACKGROUND WORK menting algorithms in standardized high-level languages [2], A. Dataflow model of computations [3], [4]. Dataflow MoCs are architecture agnostic, making them Dataflow programming models have a long and rich history highly valuable for the specification of performance portable dating back to the early 1970s [5], [6]. As depicted in Fig. 1a, a applications that can be deployed on a wide variety of com- dataflow program is defined as a (hierarchical) directed graph puting platforms. Dataflow programs are mapped to specific in which nodes (called actors) represent the computational ker- nels and directed edges (called buffers) represent the lossless, ∗This work has been done while author was with EPFL SCI STI MM order preserving, and point-to-point communication channels between actors. Buffers are used to communicate sequences need to be present for the actor to execute a step (i.e., for it to of atomic data packets (called tokens). In literature, several be enabled). For a given input sequence, the firing functions variants of dataflow Models of Computation (MoC) have been determine a sequence/state combination for which the actor is introduced [6], [7], [8]. One of their common properties is enabled according to the firing rule, the output tokens produced that individual actors encapsulate their own state which is in such step, and, if applicable, the next actor state. It must not shared among other actors of the same program. Instead, be observed that at each step only one action can be selected actors communicate with each other exclusively by sending and fired. In general, DPN actors may be non-deterministic, and receiving tokens by means of buffers connecting them. which means that the firing function may yield more than The absence of race conditions makes the behavior of dataflow one combination of outputs and next states. Furthermore, the programs more robust to different execution policies, whether execution can be totally dynamic, meaning that the number of those be truly parallel or some interleaving of the individual consumed/produced tokens may vary according to the input actors. sequence, which severely limits any compile time analysis of DPN actors. b1 b2 Producer Filter Consumer B. Dataflow programming languages (a) An example of a dataflow network with three actors (i.e. Producer, In the last decades, a plethora of different programming Filter and Consumer) and two buffers (i.e. b1 and b2). languages has been used to model and implement dataflow programs [9]. Imperative languages (e.g. C/C++, Java, Python) actor Producer() ==> i n t Y : i n t cnt := 0; have been extended with parallel constructs, or pure dataflow produce : action ==> Y : [ cnt ] languages (e.g. Ptolemy, Esterel) have been formalized. The guard cnt < 3 do RVC-CAL language [10] is the sole standardized dataflow cnt := cnt + 1; end programming language which fully captures the behavioral end features of DPN. As an example, the RVC-CAL program re- (b) Producer.cal: at each firing it produces a token on output port Y. ported in Fig. 1a is composed of three actors (i.e., Producer, Filter and Consumer) and two buffers (i.e., b1 and b2). actor Filter() int X ==> i n t Y : copy : action X : [ a ] ==> Y : [ a ] end The Producer actor (see Fig. 1b) has an output port Y i n v e r t : action X : [ a ] ==> Y:[−a ] end connected to b1, the actor internal variable cnt and the action schedule fsm state1 : state1(copy) −> state2 ; (i.e., firing function) labeled as produce. During each firing state2(invert) −> state1 ; end of the action, the value of cnt is modified and a token is end produced on b1. Furthermore, the execution of the action is (c) Filter.cal: at each firing it consumes a token from input port X and cnt Filter it produces a token on output port Y. guarded by the guard condition on . The actor (see Fig. 1c) has an input port X connected to b1, an output port actor Consumer() int X ==> : Y connected to b2, the two actions copy and invert, and an consume : action X : [ a ] ==> end end actor state machine (FSM) which drives the action selection. The selected action can be fired only if a token is available (d) Consumer.cal: at each firing it consumes a token from input port X. on b1 and there is at least one token place on b2. During each Fig. 1. RVC-CAL program example: dataflow network configuration and firing, a token is consumed from b1 and a token is produced actors source code. on b2. The Consumer actor (see Fig. 1d) is composed by the input port X connected to b2 and the action consume. The DPN is an expressive MoC in comparison with other action can be fired only if a token is available on b2. During dataflow models, e.g. by supporting dynamic branching with each firing, a token is consumed from b2. actor firings predicated on token values. This makes the DPN model sufficient for expressing dynamic, complex algorithms, C. Shared-variable approaches but comes at the cost of analysis and optimization opportuni- Conventional dataflow encapsulates isolated state inside ties. To execute, DPN actors perform a sequence of discrete actors. Their only means of communication is by sending computational steps (or firings). During each step, an actor and receiving data through dataflow edges. For example, the can consume a finite number of input tokens, produce a finite only way for the Filter and Consumer actors to know number of output tokens, and modify its own internal state the value of the cnt variable in Producer is via explicit variables if it has any.
Recommended publications
  • Design Automation for Integrated Optics
    DESIGN AUTOMATION FOR INTEGRATED OPTICS by Christopher Condrat A dissertation submitted to the faculty of The University of Utah in partial fulfillment of the requirements for the degree of Doctor of Philosophy Department of Electrical and Computer Engineering The University of Utah August 2014 Copyright © Christopher Condrat 2014 All Rights Reserved The University of Utah Graduate School STATEMENT OF DISSERTATION APPROVAL The dissertation of Christopher Condrat has been approved by the following supervisory committee members: Priyank Kalla , Chair 12 / 18 / 2013 Date Approved Steven Blair , Member 12 / 18 / 2013 Date Approved Chris Myers , Member 12 / 18 / 2013 Date Approved Kenneth Stevens , Member 12 / 18 / 2013 Date Approved Erik Brunvand , Member 12 / 18 / 2013 Date Approved and by Gianluca Lazzi , Chair/Dean of the Department/College/School of Electrical and Computer Engineering and by David B. Kieda, Dean of The Graduate School. ABSTRACT Recent breakthroughs in silicon photonics technology are enabling the integration of optical devices into silicon-based semiconductor processes. Photonics technology enables high-speed, high-bandwidth, and high-fidelity communications on the chip-scale—an important development in an increasingly communications-oriented semiconductor world. Significant developments in silicon photonic manufacturing and integration are also enabling investigations into applications beyond that of traditional telecom: sensing, filtering, signal processing, quantum technology—and even optical computing. In effect, we are now seeing a convergence of communications and computation, where the traditional roles of optics and microelectronics are becoming blurred. As the applications for opto-electronic integrated circuits (OEICs) are developed, and manufac- turing capabilities expand, design support is necessary to fully exploit the potential of this optics technology.
    [Show full text]
  • The C Programming Language
    The C programming Language The C programming Language By Brian W. Kernighan and Dennis M. Ritchie. Published by Prentice-Hall in 1988 ISBN 0-13-110362-8 (paperback) ISBN 0-13-110370-9 Contents ● Preface ● Preface to the first edition ● Introduction 1. Chapter 1: A Tutorial Introduction 1. Getting Started 2. Variables and Arithmetic Expressions 3. The for statement 4. Symbolic Constants 5. Character Input and Output 1. File Copying 2. Character Counting 3. Line Counting 4. Word Counting 6. Arrays 7. Functions 8. Arguments - Call by Value 9. Character Arrays 10. External Variables and Scope 2. Chapter 2: Types, Operators and Expressions 1. Variable Names 2. Data Types and Sizes 3. Constants 4. Declarations http://freebooks.by.ru/view/CProgrammingLanguage/kandr.html (1 of 5) [5/15/2002 10:12:59 PM] The C programming Language 5. Arithmetic Operators 6. Relational and Logical Operators 7. Type Conversions 8. Increment and Decrement Operators 9. Bitwise Operators 10. Assignment Operators and Expressions 11. Conditional Expressions 12. Precedence and Order of Evaluation 3. Chapter 3: Control Flow 1. Statements and Blocks 2. If-Else 3. Else-If 4. Switch 5. Loops - While and For 6. Loops - Do-While 7. Break and Continue 8. Goto and labels 4. Chapter 4: Functions and Program Structure 1. Basics of Functions 2. Functions Returning Non-integers 3. External Variables 4. Scope Rules 5. Header Files 6. Static Variables 7. Register Variables 8. Block Structure 9. Initialization 10. Recursion 11. The C Preprocessor 1. File Inclusion 2. Macro Substitution 3. Conditional Inclusion 5. Chapter 5: Pointers and Arrays 1.
    [Show full text]
  • Reducing State Explosion for Software Model Checking with Relaxed Mcms 3 We Do Not Distinguish the Framework from Its Implementation and Refer to Both As Mc- SPIN
    Reducing State Explosion for Software Model Checking with Relaxed Memory Consistency Models Tatsuya Abe1, Tomoharu Ugawa2, Toshiyuki Maeda1, and Kousuke Matsumoto2 1 {abet,tosh}@stair.center STAIR Lab, Chiba Institute of Technology 2 {ugawa,matsumoto}@plas.info.kochi-tech.ac.jp Kochi University of Technology Abstract. Software model checking suffers from the so-called state explosion problem, and relaxed memory consistency models even worsen this situation. What is worse, parameterizing model checking by memory consistency mod- els, that is, to make the model checker as flexible as we can supply definitions of memory consistency models as an input, intensifies state explosion. This pa- per explores specific reasons for state explosion in model checking with multi- ple memory consistency models, provides some optimizations intended to mit- igate the problem, and applies them to McSPIN, a model checker for memory consistency models that we are developing. The effects of the optimizations and the usefulness of McSPIN are demonstrated experimentally by verifying copying protocols of concurrent copying garbage collection algorithms. To the best of our knowledge, this is the first model checking of the concurrent copying protocols under relaxed memory consistency models. Keywords: software model checking; relaxed memory consistency models; state explosion; reordering of instructions; integration of states; concurrent copying garbage collection 1 Introduction Modern computing systems are based on concurrent/parallel processing designs for their performance advantages, and programs therefore must also be written to exploit these designs. However, writing such programs is quite difficult and error-prone, be- cause humans cannot exhaustively consider the behaviors of computers very well. One approach to this problem is to use software model checking, in which all possible states that can be reached during a program’sexecutionare explored.Many such model check- ers have been developed (e.g., [12,18,26,25,7,8]).
    [Show full text]
  • The C Programming Language
    The C programming Language By Brian W. Kernighan and Dennis M. Ritchie. Published by Prentice-Hall in 1988 ISBN 0-13-110362-8 (paperback) ISBN 0-13-110370-9 Contents Preface Preface to the first edition Introduction 1. Chapter 1: A Tutorial Introduction 1. Getting Started 2. Variables and Arithmetic Expressions 3. The for statement 4. Symbolic Constants 5. Character Input and Output 1. File Copying 2. Character Counting 3. Line Counting 4. Word Counting 6. Arrays 7. Functions 8. Arguments - Call by Value 9. Character Arrays 10. External Variables and Scope 2. Chapter 2: Types, Operators and Expressions 1. Variable Names 2. Data Types and Sizes 3. Constants 4. Declarations 5. Arithmetic Operators 6. Relational and Logical Operators 7. Type Conversions 8. Increment and Decrement Operators 9. Bitwise Operators 10. Assignment Operators and Expressions 11. Conditional Expressions 12. Precedence and Order of Evaluation 3. Chapter 3: Control Flow 1. Statements and Blocks 2. If-Else 3. Else-If 4. Switch 5. Loops - While and For 6. Loops - Do-While 7. Break and Continue 8. Goto and labels 4. Chapter 4: Functions and Program Structure 1. Basics of Functions 2. Functions Returning Non-integers 3. External Variables 4. Scope Rules 5. Header Files 6. Static Variables 7. Register Variables 8. Block Structure 9. Initialization 10. Recursion 11. The C Preprocessor 1. File Inclusion 2. Macro Substitution 3. Conditional Inclusion 5. Chapter 5: Pointers and Arrays 1. Pointers and Addresses 2. Pointers and Function Arguments 3. Pointers and Arrays 4. Address Arithmetic 5. Character Pointers and Functions 6. Pointer Arrays; Pointers to Pointers 7.
    [Show full text]
  • The POWER4 Processor Introduction and Tuning Guide
    Front cover The POWER4 Processor Introduction and Tuning Guide Comprehensive explanation of POWER4 performance Includes code examples and performance measurements How to get the most from the compiler Steve Behling Ron Bell Peter Farrell Holger Holthoff Frank O’Connell Will Weir ibm.com/redbooks International Technical Support Organization The POWER4 Processor Introduction and Tuning Guide November 2001 SG24-7041-00 Take Note! Before using this information and the product it supports, be sure to read the general information in “Special notices” on page 175. First Edition (November 2001) This edition applies to AIX 5L for POWER Version 5.1 (program number 5765-E61), XL Fortran Version 7.1.1 (5765-C10 and 5765-C11) and subsequent releases running on an IBM ^ pSeries POWER4-based server. Unless otherwise noted, all performance values mentioned in this document were measured on a 1.1 GHz machine, then normalized to 1.3 GHz. Note: This book is based on a pre-GA version of a product and may not apply when the product becomes generally available. We recommend that you consult the product documentation or follow-on versions of this redbook for more current information. Comments may be addressed to: IBM Corporation, International Technical Support Organization Dept. JN9B Building 003 Internal Zip 2834 11400 Burnet Road Austin, Texas 78758-3493 When you send information to IBM, you grant IBM a non-exclusive right to use or distribute the information in any way it believes appropriate without incurring any obligation to you. © Copyright International Business Machines Corporation 2001. All rights reserved. Note to U.S Government Users – Documentation related to restricted rights – Use, duplication or disclosure is subject to restrictions set forth in GSA ADP Schedule Contract with IBM Corp.
    [Show full text]
  • Accelerating Applications at Scale Using One-Sided Communication
    Accelerating Applications at Scale Using One-Sided Communication Hongzhang Shan, Brian Austin, Nicholas J. Wright, Erich Strohmaier, John Shalf and Katherine Yelick CRD and NERSC Lawrence Berkeley National Laboratory, Berkeley, CA 94720 hshan, baustin, njwright, estrohmaier, jshalf, [email protected] Abstract—The lower overhead associated with one-sided mes- their use in this paper. The principle reason is that performance saging as compared to traditional two-sided communication has measurements on our experimental platform show that MPI-2 the potential to increase the performance at scale of scientific one-sided performs significantly worse than PGAS and in fact applications by increasing the effective network bandwidth and reducing synchronization overheads. In this work, we investigate worse than the MPI two-sided. both the programming effort required to modify existing MPI CAF and UPC have both been around for a decade or so. applications to exploit one-sided messaging using PGAS lan- Neither of them however has been widely adopted by the guages, and the resulting performance implications. In particular, user community; partly because of the lack of a developer we modify the MILC and IMPACT-T applications to use the environment, and partly because of the lack of convincing one-sided communication features of Unified Parallel C (UPC) and coarray Fortran (CAF) languages respectively. Only modest performance results for real applications, especially at large- modifications to the source code are required where fewer than scale, that demonstrate why they are viable (or superior) 100 lines of the source code out of 70,000 need to be changed for alternatives to the MPI programming model.
    [Show full text]
  • APLSV User's Manual APL SHARED VARIABLE SYSTEM
    APLSV User's Manual APL SHARED VARIABLE SYSTEM ~-""L r. l. LL L LL ':' L A A A A A S5'~"'''' 'V~V¥LLLLL!LLLLAAAAA SSS~SSS$SS PPPPPLLLLLYVVVV VYVVVLLLLLLLLLLAAAAA ssssssssss PPPPPLLLLLVYVVV LLLLLSSSS$ PPPPP PPPPPssssssssss VVVYVLLLLL LLLLLSSSSS PPPPP PPPPPSSSSSS5SSS VVVVVLLLLL LLLLLSSSSS PPPPP rrrrrssssssssss VVVVVLLLLL SSSSPFPPPS5SSS VVYVYAAAAA VVVYVPPPPPsssssvvvvv pppp 5SSSPPPPPSSS$S VVVVVAAHA VVVVI'PPPPPSSSSSVVVYV PPE'P SsssSPPPPPSSS$S VVVVVAAAAA ~VVVYPPPPPSSSSSVVVYV pppp AAAA VVVVVVVVVVSSSSS SBSSS AAAAALLLLLVVYVVLLL'LAAAA ,.AHA vvvvvvrvvvsssss esss: AAAAALLLLLVVVVVL LL HAd ~AAAA VVVVVVYVVVSSSSS s.:~~ AAAAALLLLLVVVVVLLLLLAAAA HA"?PP?PAAAA"!.LLLL [,!.LLLLLLLLPPPPP .. :'L;' .lAAAAPPPPPAAAAALLLn LLLLLLLLLLPPPPI' LL'· AAAAPPPPf'AAAAALULL LLLL' r.LLLLPPPPP "LL:' AAAA LLLLLPPPPPAAAAA LJ..LL ULLL 1'1" v AAAA LLLLLl'PPl'PAAAAA LLLLLLLLL! ilVVV AAA.A LLLLLPPPPPA.AA.AA tLLLLLLLLL VHV VVVVVV7VV SSSSS SSSSSYVVVYYYVVY AAAAAPPPP VVVVVYYVV SSSSS SSSSSVYVVYYVVYY AAAAAPPPP VYYVVVYVY SSSSS SSSSSVVYVVVVVVV AAAAAPPPP PPPPLLLLLSSSSSVYVYVPPPPi'VVVVVAAAAAi'Pi'PP[,LLLLLLLLLPPFPPAAAAA~AAA PPPPLLLLL SSSSYVVVVPPi'PPVVVYYAAAAAPPpPpLLLLLLLLLLPPPPPAAAAAAAAA pPPI'L['LLLSSSSSVYVVVI'PPPPVYVVVAA.AA.A.PPPPPLL~~LLL·LLPPPPPA.AA.AAAA.AA. SSSSYVYVY LLLLLLLL!L vvvrv 'LL' SSSSSVVVYVAAAAA.AA.AA SSSSYYV'. LLLLLLLLLL VVVYV~L'LL JSSSSVYVYVAAAAAAAAA SSSSVYJlVV LLLLLLLLLL JlJIV"VL':'':'' L _SSSSYVVYVAAAAAAAA.A YVVVV AAAAA PPPPP SSSS AAAAAAAAAAPPPPP VVVY AAAAA PPPPP SS5SS AAAAAAAAAAP PPE' VVYVV AAAAA PPPPP SSSSS AAAAAAAAAAPPPPP SSSSAAAAASSSSSPPPPPPPPPPLLLLLAAAAASSSSSYVYVV$SSSS LLLLLAAAA
    [Show full text]
  • The C Programming Language
    THE C PROGRAMMING LANGUAGE "'YC{9//6W ( BWf ) Library of Congress Calaloging in Publicolion Dato KERNICHAN, BRIAN W. The C programming language. lncludes index. l. C (Computer program language) l. RITCHIE, DENNIS M., joint author. ll. Title. QA76.73.C!5K47 Oot.6'424 77-28983 ISBN 0-13-t 10163-3 Copyright @ 1978 by Bell Telephone Laboratories, Incorporated. All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form or by any means, electronic, mechanical, photocopy- ing, recording, or otherwise, without the prior written permission of the publisher. Print- ed in the United States of America. Published simultaneously in Canada. This book was set in Times Roman and Courier l2 by the authors, using a Graphic Sys- tems phototypesetter driven by a PDP-l l/70 running under the UNIX operating system. UNIX is a Trademark of Bell Laboratories. l5 14 PRENTICE-HALL INTERNATIONAL, INC., London PRENTICE-HALL OF AUSTRALIA PTY. LIMITED, Sydney PRENTICE-HALL OF CANADA, LTD., Toronto PRENTICE-HALL OF INDIA PRIVATE LIMITED, New Delhi PRENTICE-HALL OF JAPAN, INC., Tokyo PRENTICE-HALL OF SOUTHEAST ASIA PTE. LTD., Singapore WHITEHALL BOOKS LIMITED, Wellington, New Zealand CONIENTS Preface lx Chapter 0 Introduction Chapter I A Tutorial Introduction 1.1 Getting Started 5 1.2 Variables and Arithmetic 8 1.3 The For Statement ll t.4 Symbolic Constants t2 1.5 A Collection of Useful Programs l3 1.6 Arrays 20 t.7 Functions 22 1.8 Arguments - Call by Value 24 1.9 Character Arrays 25 l.l0 Scope; External
    [Show full text]
  • APLX for Windows, Linux and Macos
    APLX for Windows, Linux and MacOS New features in Version 2.0 Copyright © 2003 MicroAPL Ltd. All rights reserved worldwide. APLX, APL.68000 and MicroAPL are trademarks of MicroAPL Ltd. All other trademarks acknowledged. APLX is a proprietary product of MicroAPL Ltd, and its use is subject to the license agreement in force. Unauthorized copying or use of APLX is illegal. MicroAPL Ltd makes no warranties in respect of the suitability of APLX for any particular purpose, and accepts no liability for any loss arising out of the use of APLX or arising from the information contained in this manual. MicroAPL welcomes your comments and suggestions. Please visit our website: http://www.microapl.co.uk/apl APLX Version 2.0.3 Upgrade notes: February 2004 2 Summary of APLX Version 2 Enhancements Multi-tasking • Create a new session from the menu, creates an independent APL task with its own session window • Create a new APL task under program control, optionally with session window: 'ChildTask' ŒWI 'New' 'APL' ('wssize' 100000) ('visible' 1) • New properties and methods for controlling APL child tasks, eg Execute method causes child task to execute APL expression or system command • New events allow parent task to be notified when child task starts or ends execution of an expression, hits an untrapped error, or terminates • Signals for the child and parent task to talk to each other and pass commands/results • Tasks can share data through shared variables and ŒWI delta-properties Interpreter • Removal of 64KB limit on executing token strings • Increase
    [Show full text]
  • Programming Language) 1 APL (Programming Language)
    APL (programming language) 1 APL (programming language) APL Paradigm array, functional, structured, modular Appeared in 1964 Designed by Kenneth E. Iverson Developer Kenneth E. Iverson Typing discipline dynamic Major implementations IBM APL2, Dyalog APL, APL2000, Sharp APL, APLX Dialects A+, Dyalog APL, APLNext Influenced by mathematical notation [1] [2] [3] [4] Influenced J, K, Mathematica, MATLAB, Nial, PPL, Q APL (named after the book A Programming Language)[5] is an interactive array-oriented language and integrated development environment which is available from a number of commercial and non-commercial vendors[6] and for most computer platforms.[7] It is based on a mathematical notation developed by Kenneth E. Iverson. APL has a combination of unique and relatively uncommon features that appeal to programmers and make it a productive programming language:[8] • It is concise, using symbols rather than words and applying functions to entire arrays without using explicit loops. • It is solution focused, emphasizing the expression of algorithms independently of machine architecture or operating system. • It has just one simple, consistent, and recursive precedence rule: the right argument of a function is the result of the entire expression to its right. • It facilitates problem solving at a high level of abstraction. APL is used in scientific,[9] actuarial,[8] statistical,[10] and financial applications where it is used by practitioners for their own work and by programmers to develop commercial applications. It was an important influence on the development of spreadsheets, functional programming,[11] and computer math packages.[3] It has also inspired several other programming languages.[1] [2] [4] It is also associated with rapid and lightweight development projects in volatile business environments.[12] History The first incarnation of what was later to be the APL programming language was published and formalized in A Programming Language,[5] a book describing a notation invented in 1957 by Kenneth E.
    [Show full text]
  • Shared Variables Interaction Diagrams
    Shared Variables Interaction Diagrams Rajeev Alur Radu Grosu Department of Computer and Information Science Department of Computer Science University of Pennsylvania State University of New York at Stony Brook [email protected] [email protected] Abstract variables paradigm. Textbooks on concurrent programming (e.g., [14, 17]) contain many pictures describing the interac- Scenario-based specifications offer an intuitive and vi- tions of processes communicating by shared variables, and sual way of describing design requirements of distributed similar scenarios arise in diverse areas such as transaction software systems. For the communication paradigm based processing in concurrent databases (c.f. [18]) and consis- on messages, message sequence charts (MSC) offer a stan- tency in shared-memory multiprocessors [13]. dardized and formal notation amenable to formal analysis. In our definition of an SVID, an action corresponds to, In this paper, we define shared variables interaction dia- possibly multiple, reading/writing of shared variables. The grams (SVID) as the counterpart of MSCs when processes actions of one process are visually ordered. The causal de- communicate via shared variables. After formally defining pendence among actions of different processes is illustrated SVIDs, we develop an intuitive as well as formal definition Ô by arrows: an arrow from an action a of process to an ac- of refinement for SVIDs. This notion provides a basis for Õ Õ tion b of process means that reads a value that was writ- systematically adding details to SVID requirements. ten by Ô. If the variable involved in this communication is write-shared, then there is an implicit additional constraint b that between these two actions a and , there is no interven- 1.
    [Show full text]
  • INFORMATION to USERS This Manuscript Has Been Reproduced
    INFORMATION TO USERS This manuscript has been reproduced from the microfilm master. UMI films the text directly from the original or copy submitted. Thus, some thesis and dissertation copies are in typewriter face, while others may be from any type of computer printer. The quality of this reproduction is dependent upon the quality of the copy submitted. Broken or indistinct print, colored or poor quality illustrations and photographs, print bleedthrough, substandard margins, and improper alignment can adversely affect reproduction. In the unlikely event that the author did not send UMI a complete manuscript and there are missing pages, these will be noted. Also, if unauthorized copyright material had to be removed, a note will indicate the deletion. Oversize materials (e.g., maps, drawings, charts) are reproduced by sectioning the original, beginning at the upper left-hand corner and continuing from left to right in equal sections with small overlaps. Each original is also photographed in one exposure and is included in reduced form at the back of the book. Photographs included in the original manuscript have been reproduced xerographically in this copy. Higher quality 6" x 9" black and white photographic prints are available for any photographs or illustrations appearing in this copy for an additional charge. Contact UMI directly to order. A Bell & Howell Information Com pany 300 North Zeeb Road. Ann Arbor, Ml 48106-1346 USA 313/761-4700 800/521-0600 A l g o r it h m s t o I m p l e m e n t S e m a p h o r e s in D is t r ib u t e d E nvironments dissertation Presented in Partial Fulfillment of the Requirements for the Degree Doctor of Philosophy in the Graduate School of The Ohio State University By Mahendra A.
    [Show full text]