Lecture Notes in Computer Science Vol 191, Springer-Verlag (1985)

Total Page:16

File Type:pdf, Size:1020Kb

Lecture Notes in Computer Science Vol 191, Springer-Verlag (1985) ACKNOWLEDGEMENT We are grateful to Ian Pyle for his valuable comments on a draft of this book. APPENDIX A Task Initialisation in Highly Parallel Systems In many applications it is necessary for a task to know its identity so that, for example, resource allocation mechanisms can be implemented reliably. Unfortunately as an Ada task cannot be initialised at the time of creation, a rendezvous must be used to pass identification data to those tasks that require it. With multiprocessor systems this initialisation phase can be a source of inefficiency as it is, essentially, a sequential activity. In this appendix we present Bums' algorithm 42 for initialising tasks in highly parallel systems. Each task is presented in an initial rendezvous with a range (B..T) of identities of uninitialised tasks. The task initialises its own identity to the top element (T) of the range. Following the rendezvous, the task splits its given range in two and passes the top half to another uninitialised task. It then repeats the process on the bottom half; i.e. it splits it in two, passes the top half on, and so on. Eventually the bottom half will consist of a single identity (top = base) and, having passed this identity on, the task will have completed its initialisation role. It is easy to see that the time taken to initialise all the tasks depends logarithmically on the number of tasks. Thus over ten thousand tasks can be initialised within 14 rendezvous times. 108 A REVIEW OF ADA TASKLNG procedure main is max : constant := ~oo; --number of tasks to be created subtype task_range is integer range 0 .. max; task type TK is entry initialise(B,T : task_range); end TK; set : array(task_range) of TK; base, new_base : task_range; top : integer range task_range~first-! .o task_range'last; task body TK is my_number : task_range; -- other declarations begin declare base,new_base : task_range; top : integer range task_range'first-i oo task_range'last; begin accept initialise(B,T : task_range) do my_number := T; base := B; end initialise; top := my_number - i; while top >= base loop new_base :: (top + base)/2; set(top).initialise(new_base,top); --top half of remaining group top := new_base - I; end loop; end; end TK; APPENDIX A 109 begin top :: task_range'last; base := task_range'first; while top >= base loop new_base := (top + base)/2; set(top).initialise(new_base,top); --top half of remaining group top := new_base - i; end loop; end main; APPENDIX B Asynchronous Agents Normal inter-task communication is synchronous. This appendix contains generic packages to implement asynchronous send and receive operations. It also gives an examples of their use. First the asynchronous send: generic type mess is private; with procedure destination(m : mess); package no_wait_send is task type agent is entry forward(m : mess); end agent; type agt is access agent; end no_wait_send; package body no_wait_send is task body agent is temp : mess; begin accept forward(m : mess) do temp := m; end; destination(temp); end agent; end no_wait_send; The generic package provides a task which stores a single message (for the client that created it) and passes on this message by calling the procedure destination which is 112 A REVIEW OF ADA TASKING provided as a generic parameter. A simple example of use is a main program calling (indirectly) the entry 'request' in the task 'server': with no_wait_send; procedure main is type message is new integer; task server is entry request(m : message); end server; package messenger is new no_wait_send(mess => message, destination => server.request); use messenger; phidippides : agz := new agent; m : message; task body server is separate; begin -- obtain value for m phidippides.forward(m); -- continue execution end main; The following package implements an asynchronous receive. APPENDIX B 113 generic type mess is private; package mailbox is task type agent is entry deposit(m : mess); entry collect(m : out mess); end agent; type agt is access agent; end mailbox; package body mailbox is task body agent is temp : mess; begin accept deposit(m : mess) do temp := m; end; accept collect(m : out mess) do m := temp; end; end agent; end mailbox; With this generic package a task type is provided that has two entry calls, one for the message to be deposited and the other to allow the client task to collect the message when it is ready to do so. Use in a server package is illustrated below. serverl ' call "~-Ic°ntr°ll Iag ent ] A server task will call the procedure transaction in order to pass data to control and receive the results. The transaction procedure, however, calls up a mailbox agent to handle the return data. The control task therefore receives the input data and an "address" (access variable) to where the reply should be sent. 114 A REVIEW OF ADA TASKING package server is type input_data is new integer; type output_data is new integer; procedure transaction(I :input_data; 0 : out output_data); end server; with mailbox; package body server is package mbox is new mailbox(mess => output_data ; use mbox; task control is entry call(I : input_data; A : agt); end control; procedure transaction(I:input_data;0:out output_data) is A : agt := new agent; begin control.call(I,A); -- give input data to the -- control task together with -- a mailbox agent for the -- return data A.coi!ect(O) ; -- wait for data to be -- ready in mailbox end transaction; task body control is temp_I : input_data; temp_0 : output_data; temp_A : agt; begin accept call(I : input_dana; A : agt) do temp_I := I; temp_A := A; end; -- produce temp_0 from temp_I temp_A.deposit(temp_0); -- when appropriate end control; end server; APPENDIX C Implementation of a Reliable Transaction This appendix contains a generic package that allows a reliable two-rendezvous transaction to be programmed in Ada. The reliability comes from two features: (a) The server task returns data to the client (during the second rendezvous) via an agent. The failure or abortion of the client does not therefore affect the server. (b) A pool of reusable agents is used to overcome the storage exhaustion problem (see section 4.1.2). The generic package has the specification shown below. The generic parameters provide a type for the re'turn data and a constant giving the size of the pool. For simplicity, this example uses a fixed size pool of agents. A dynamic pool is also possible, in which case the constant parameter would give the maximum size of the pool. 116 A REVIEW OF ADA TASKING generic type mess is private; size_of_pool : integer := 16; package agent_pool is type agent is private; procedure deposit(agt : agent; m : mess); procedure collect(agt : agent; m : out mess); procedure allocate(agt : out agent); procedure free(agt : in out agent); NO_AVAILABLE_AGENT : exception; -- raised by allocate private task type agent_t is entry deposit(m : mess); entry collect(m : out mess); end agent_t; type agent_p is access agent_t; type agent is record in_use : boolean :: FALSE; ptr : agent_p := new agent_t; end record; end agent_pool; The generic package provides four procedures: two for obtaining and releasing an agent, and two for communication with the agent. An agent can therefore be used a number of times by a client before being returned to the pool. The agent itself has the form of an asynchronous receiver (see Appendix B). Also defined within the package is an exception that is raised if there is no free agent available. The body for the generic package is as follows. (Note that all agents and the agent manager terminate when required to do so.) APPENDIX C 117 package body agent_pool is task manager is entry allocate(agt : out agent); e~ry free(agt: in out agent); end manager; procedure deposit(agt : agent; m : mess) is begin agt.ptr.deposit(m); end deposit; procedure collect(agt : agent; m : out mess) is begin agt.ptr.collect(m); end collect; procedure allocate(agt : out agent) is begin manager.allocate(agt); end allocate; procedure free(agt : in out agent) is begin manager.free(agt); end free; task body agent_t is temp : mess; begin loop select accept deposit(m : mess) do temp := m; end; or terminate; end select; accept collect(m : out mess) do m := temp; end; end loop; end agent_t; ]8 A REVIEW OF ADA TASK~G task body manager is subtype pool_size is integer range i.. size_of_pool; agent_pool : array (pool_size) of agent; found : boolean begin loop loop -- first all outstanding calls -- to free are processed select accept free(agt : in out agent) do agt.in_use := FALSE; end free; else exit; end select; end loop; begin select accept alloeate(agt : out agent) do found := FALSE; for i in pool_size loop if not agent_pool(i).in_use then agent_pool(i).in_use :: TRUE; agt := agent_pool(i); found := TRUE; exit; end if; end loop; if not found then raise NO_AVAILABLE_AGENT; end if; end allocate; or accept free(agt : in out agent) do agt.in_use := FALSE; end free; or terminate; end select; exception when NO_AVAILABLE_AGENT => null; end; end loop; end manager; end agent_pool; APPENDIX C 119 To illustrate the use of this generic package the simple server task employed in Appendix B is implemented (reliably!): package server is type input_data is new integer; type output_data is new integer; procedure transaction(I:input_data;0:out output_data); end server; with agent_pool; package body server is package reliable_agents is new agent_pool(mess => output_data); task control is entry call(I:input_data;A:reliable_agents.agent); end control; procedure transaction(I:input_data;0:out output_data) is A :reliable_agents.agent; begin reliable_agents.allocate(A); control.call(I,A); reliable_agents.collect(A,O); reliable_agents.free(A); end transaction; task body control is temp_l : input_data; temp_0 : output_data; temp_A : reliable_agents.agent; begin accept call(I:input_data;A:re!iable_agents.agent) do temp_I := I; temp_A := A; end; reliable_agents.deposit(temp_A, temp_0);--when appropriate end control; end server; APPENDIX D Using Families For Handling Priority Requests In Section 4.2.6.3 a prioritised select was implemented to give preference to one entry over another in a two accept select statement.
Recommended publications
  • Concurrent Programming Introduction
    v1.0 20130407 Programmazione Avanzata e Paradigmi Ingegneria e Scienze Informatiche - UNIBO a.a 2013/2014 Lecturer: Alessandro Ricci [module 2.1] CONCURRENT PROGRAMMING INTRODUCTION PAP LM - ISI - Cesena - UNIBO !1 SUMMARY • Concurrent programming – motivations: HW evolution – basic jargon • processes interaction, cooperation, competition, • mutual exclusion, synchronization • problems: deadlocks, starvation, livelocks • A little bit of history – Dijkstra, Hoare, Brinch-Hansen • Concurrent languages, mechanisms, abstractions – overview PAP LM - ISI - Cesena - UNIBO Introduction !2 CONCURRENCY AND CONCURRENT SYSTEMS • Concurrency as a main concept of many domains and systems – operating systems, multi-threaded and multi-process programs, distributed systems, control systems, real-time systems,... • General definitions – “In computer science, concurrency is a property of systems in which several computational processes are executing at the same time, and potentially interacting with each other.” [ROS-97] – “Concurrency is concerned with the fundamental aspects of systems of multiple, simultaneously active computing agents, that interact with one another” [CLE-96] • Common aspects – systems with multiple activities or processes whose execution overlaps in time – activities can have some kind of dependencies, therefore can interact PAP LM - ISI - Cesena - UNIBO Introduction !3 CONCURRENT PROGRAMMING • Concurrent programming – building programs in which multiple computational activities overlap in time and typically interact in some way • Concurrent program – finite set of sequential programs that can be executed in parallel, i.e. overlapped in time • a sequential program specifies sequential execution of a list of statements • the execution of a sequential program is called process • a concurrent program specifies two or more sequential programs that may be executed concurrently as parallel processes – the execution of a concurrent program is called concurrent computation or elaboration PAP LM - ISI - Cesena - UNIBO Introduction !4 CONCURRENT PROGRAMMING VS.
    [Show full text]
  • The Programming Language Concurrent Pascal
    IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, VOL. SE-I, No.2, JUNE 1975 199 The Programming Language Concurrent Pascal PER BRINCH HANSEN Abstract-The paper describes a new programming language Disk buffer for structured programming of computer operating systems. It e.lt­ tends the sequential programming language Pascal with concurx:~t programming tools called processes and monitors. Section I eltplains these concepts informally by means of pictures illustrating a hier­ archical design of a simple spooling system. Section II uses the same enmple to introduce the language notation. The main contribu~on of Concurrent Pascal is to extend the monitor concept with an .ex­ Producer process Consumer process plicit hierarchy Of access' rights to shared data structures that can Fig. 1. Process communication. be stated in the program text and checked by a compiler. Index Terms-Abstract data types, access rights, classes, con­ current processes, concurrent programming languages, hierarchical operating systems, monitors, scheduling, structured multiprogram­ ming. Access rights Private data Sequential 1. THE PURPOSE OF CONCURRENT PASCAL program A. Background Fig. 2. Process. INCE 1972 I have been working on a new programming .. language for structured programming of computer S The next picture shows a process component in more operating systems. This language is called Concurrent detail (Fig. 2). Pascal. It extends the sequential programming language A process consists of a private data structure and a Pascal with concurrent programming tools called processes sequential program that can operate on the data. One and monitors [1J-[3]' process cannot operate on the private data of another This is an informal description of Concurrent Pascal.
    [Show full text]
  • Concurrent Pascal
    LIBP.ARY CLEARANCE 1. (a) I give my permission for my thesis, entitled .....~D.~~ r.r-.m1 .... P.C?.~~.: .. .) Q?y).~-tJ.\.1Ati~0.. A,:&. U~oJe- .................................................................... .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. to be made available to readers in the Library under the conditions determined by the Librarian. (b) I agree to my thesis, if asked for by another institution, being sent away on temporary loan under conditions determined by the Librarian. (e) I also agree that oy thesis may be copied for Library use. 2. I do not wish m . .. .. .. .. .. .. ......................................... ...................~--- ~-........................ .. .. .. .. .. .. ~ ................... ~- ...................... •. '" -, to be made available to re~~,or to be sent ~ll.er ins-ti-tut.ions without my written consent withiri'--the next two years~ Signed .. 1?~~-............... Date ... ?!.? I 9./.$~.................. Strike out the se»tence or phrase which does not apply. The Library Massey University Palmerston North, N.Z. The copyright of this thesis belongs to the author. Readers must ~ign their name in the space below to show that they recognise this. They are asked to add their permanent address. Name and Address Date ................................................................................. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ••••••••••••••o•••••••••••• .. •••••••••••••••••••••••••••••••••••••••••••••••••••• ................................................................................
    [Show full text]
  • Pascal Versus C : a Subjective Comparison
    PROCEEDINGS OF THE SYMPOSIUM ON LANGUAGE DESIGN AND PROGRAMMING METHODOLOGY SYDNEY, 10-11 SEPTEMBER~ 1979 PASCAL VERSUS C A SUBJECTIVE COMPARISON Prabhaker Mateti Department of Computer Science University of Melbourne ABSTRACT The two programming languages Pascal and C are subjectively compared. While the two languages have comparable data and control structures, the program structure of C appears superior. However, C has many potentially dangerous features, and requires great caution from its programmers. Other psychological effects that the various structures in these languages have on the process of programming are also conjectured. "At first sight, the idea of any rules or principles being superimposed on the creative mind seems more likely to hinder than to help, but this is really quite untrue in practice. Disciplined thinking focusses inspiration rather than blinkers it." - G. L. Glegg, The Design of Design. I Introduction Pascal has become one of the most widely accepted languages for the teaching of programming. It is also one of the most thoroughly studied languages. Several large programs have been written in Pascal and its derivatives. The programming language C has gained much prominence in recent years. The successful Unix operating system and most of its associated software are written in C. This paper confines itself to a subjective comparison of the two languages, and conjectures about the effect various structures of the languages have on the way one programs. While we do occasionally refer to the various extensions and compilers of the languages, the comparison is between the languages as they are now, and in the context of general programming.
    [Show full text]
  • BACI Pascal Compiler User Guide
    USER'S GUIDE BACI Pascal Compiler and Concurrent PCODE Interpreter Bill Bynum/Tracy Camp College of William and Mary/Colorado School of Mines July 14, 2003 BACI Pascal Compiler User's Guide 1 Contents 1 Introduction 2 2 Pascal Compiler Syntax 2 3 Concurrency Constructs 4 3.1 COBEGIN—COEND . 4 3.2 Semaphores . 4 3.2.1 Initializing a Semaphore . 4 3.2.2 P (or WAIT) and V (or SIGNAL) Procedures . 5 3.2.3 Examples of Semaphore Usage . 6 3.3 Monitors . 7 3.3.1 Condition Variables . 7 3.3.2 WAITC and SIGNALC Procedures . 7 3.3.3 Immediate Resumption Requirement . 8 3.3.4 An Example of a Monitor . 8 3.4 Other Concurrency Constructs . 8 3.4.1 ATOMIC Keyword . 9 3.4.2 PROCEDURE suspend; . 9 3.4.3 PROCEDURE revive( process number : INTEGER); . 9 3.4.4 FUNCTION which proc : INTEGER; . 9 3.4.5 FUNCTION random( range : INTEGER ): INTEGER; . 9 4 Built-in String Handling Functions 10 4.1 PROCEDURE stringCopy(dest : STRING; src : STRING); . 10 4.2 PROCEDURE stringConcat(dest : STRING; src : STRING); . 10 4.3 FUNCTION stringCompare(x : STRING; y : STRING): INTEGER; . 10 4.4 FUNCTION stringLength(x : STRING): INTEGER; . 10 4.5 FUNCTION sscanf(x : STRING; fmt : RAWSTRING ; ...): INTEGER; . 10 4.6 PROCEDURE sprintf(x : STRING; x, fmt : RAWSTRING,...); . 11 5 Using the BACI Pascal Compiler and PCODE Interpreter 11 6 Sample program and output 12 BACI Pascal Compiler User's Guide 2 1 Introduction The purpose of this document is to provide a brief description of the BACI Pascal Compiler and Concur- rent PCODE Interpreter programs and a description of how to use them.
    [Show full text]
  • Işletim Sistemleri
    Bilgisayar İŞLETİM SİSTEMLERİ Bilgisayar İŞLETİM SİSTEMLERİ Prof.Dr. Ali SAATÇİ Hacettepe Üniversitesi Bilgisayar Mühendisliği Bölümü Genişletilmiş İkinci Baskı Ankara, 2002 DEFNE ve YUNUS'a Yazar Hakkında Dr. Ali SAATÇİ, 1950 yılında Ankara'da doğdu. Orta öğrenimini Galatasaray Lisesinde tamamladı. Fransız Hükümetinin verdiği bursla Fransa'ya gitti. Institut National des Sciences Appliquées de Lyon adlı okuldan Elektrik Yüksek Mühendisliği diploması aldı. Aynı okulda, bilgisayarlı süreç denetimi konusunda doktora çalışmalarını tamam- layarak 1980 yılında yurda döndü. 1981 yılından başlayarak Hacettepe Üniversitesi, Bilgisayar Mühendisliği Bölümünde çalışmaya başladı. Bilgisayar Yapısı, Sistem Programlama, İşletim Sistemleri, Veri İletişimi, Bilgisayar Ağları, Ağ Güvenliği, Mikroişleyiciler konularında dersler verdi. Bilgisayar Donanımı anabilim dalında 1983'te Doçent, 1989 yılında da Profesör ünvanlarını aldı. 1986-1991 yılları arasında, Hacettepe Üniversitesi, Bilgisayar Mühendisliği Bölümünde Bölüm Başkanlığı yaptı. Dr. Ali SAATÇİ, halen aynı bölümde öğretim üyeliği ve Bilgisayar Donanımı Anabilim Dalı Başkanlığı görevlerini yürütmektedir. LER İ NDEK İ Ç İ ÖNSÖZ xiii GİRİŞ 1 1.1. İşletim Sisteminin Genel Tanımı ve Konumu.........................................................2 1.2. Bilgisayar Yapısının Gösteriminde Kullanılan Model............................................4 1.3. İşin Tanımı ..............................................................................................................8 1.4. Tek İş Düzeni, Çok İş Düzeni
    [Show full text]
  • The Solo Operating System the Solo Operating System
    PER BRINCH HANSEN Information Science California Institute of Technology July 1975 THE SOLO OPERATING SYSTEM THE SOLO OPERATING SYSTEM Per Brinch Hansen Information Sciencs California Institute of Technology Pasadena, California 91125 July 1975 This is a description of the Solo operating system written in the programming language Concurrent Pascal for the PDP 11/45 computer. It contains four papers entitled 1. The Solo operating system: A Concurrent Pascal program 2. The Solo operating system. Job interface 3. The Solo operating system: Processes, monitors, and classes 4, Disk scheduling at compile time The development of Concurrent Pascal and Solo has been supported by the National Science Foundation under grant number DCR74-17331 Copyright @ 1975 Par Brinch Hansen THE SOLO OPERATING SYSTEIYI. A CONCURRENT PASCAL PROGRAM Per Brinch Hansen Information Science California Institute of Technology Pasadena, California 91125 June 1975 Summary This is a description of the single-user operating system Solo written in the programming language Concurrent Pascal. It supports the development of sequential and concurrent Pascal programs for the PDP 11/45 computer. Input/output are handled by concurrent processes, Pascal programs can call one another recursively and pass arbitrary parameters among themselves. This makes it possible to use Pascal as a job control language. Solo is the first major example of a hierarchical concurrent program implemented in terms of abstract data types (classes, monitors, and processes) with compile-time control of most access rights. It is described here from the user's point of view as an introduction to anothsr paper describing its internal structure. Key Words and Phrases: Solo operating system, Concurrent Pascal, Job control, Input/output buffering, File system, Disk allocation, Operator communication, Size and performance.
    [Show full text]
  • BACI C−− Compiler and Concurrent PCODE Interpreter
    USER'S GUIDE BACI C−− Compiler and Concurrent PCODE Interpreter Bill Bynum/Tracy Camp College of William and Mary/Colorado School of Mines July 14, 2003 BACI C−− User's Guide 1 Contents 1 Introduction 2 2 C−− Compiler Syntax 2 3 Concurrency Constructs 4 3.1 cobegin Block . 4 3.2 Semaphores . 4 3.2.1 Initializing a Semaphore . 4 3.2.2 p (or wait) and v (or signal) Functions . 5 3.2.3 Examples of Semaphore Usage . 6 3.3 Monitors . 7 3.3.1 Condition Variables . 7 3.3.2 waitc and signalc Functions . 7 3.3.3 Immediate Resumption Requirement . 8 3.3.4 An Example of a Monitor . 8 3.4 Other Concurrency Constructs . 8 3.4.1 atomic Keyword . 9 3.4.2 void suspend( void ); . 9 3.4.3 void revive( int process number ); . 9 3.4.4 int which proc( void ); . 9 3.4.5 int random( int range ); . 9 4 Built-in String Handling Functions 9 4.1 void stringCopy(string dest, string src); . 9 4.2 void stringConcat(string dest, string src); . 10 4.3 int stringCompare(string x, string y); . 10 4.4 int stringLength(string x); . 10 4.5 int sscanf(string x, rawstring fmt,...); . 10 4.6 void sprintf(string x, rawstring fmt,...); . 10 5 Using the BACI C−− Compiler and PCODE Interpreter 11 6 Sample program and output 11 BACI C−− User's Guide 2 1 Introduction The purpose of this document is to provide a brief description of the C−− BACI Compiler and Concurrent PCODE Interpreter programs and a description of how to use them.
    [Show full text]
  • "Joyce Performance on a Multiprocessor" (1988)
    Syracuse University SURFACE Electrical Engineering and Computer Science - Technical Reports College of Engineering and Computer Science 9-1988 Joyce Performance on a Multiprocessor Per Brinch Hansen Syracuse University, School of Computer and Information Science, [email protected] Rangachari Anand Syracuse University, School of Computer and Information Science Follow this and additional works at: https://surface.syr.edu/eecs_techreports Part of the Computer Sciences Commons Recommended Citation Hansen, Per Brinch and Anand, Rangachari, "Joyce Performance on a Multiprocessor" (1988). Electrical Engineering and Computer Science - Technical Reports. 28. https://surface.syr.edu/eecs_techreports/28 This Report is brought to you for free and open access by the College of Engineering and Computer Science at SURFACE. It has been accepted for inclusion in Electrical Engineering and Computer Science - Technical Reports by an authorized administrator of SURFACE. For more information, please contact [email protected]. CIS-88-5 JOYCE PERFORMANCE ON A MULTIPROCESSOR Per Brinch Hansen and Anand Rangachari School of Computer & Information Science Syracuse University Syracuse, New York 13244, U.S.A. SCHOOL OF COMPUTER AND INFORMATION SCIENCE ~j PERFORMANCE ~ A ~~PROCESSOB PER BRINCH HANSEN AND ANAND RANGACHARI School of Computer and Information Science Syracuse University Syracuse, New York 13244 September 1988 AbstrAQ~ - Joyce is a parallel programming language based on CSP and Pascal. The language has been moved from the IBM PC to the Encore Multimax. The paper explains how the multiprocessor implementation of Joyce was guided by performance evaluation. The measurements show that the speed-up or Joyce programs follows Amdahl's law. lndex Ierms - Programming languages, concurrent program­ ming, communicating agents, multiprocessors, language implementation.
    [Show full text]
  • Concurrent Pascal Machine Concurrent Pascal Machine
    PER BRINCH HANSEN Information Science California Institute of Technology October 1975 CONCURRENT PASCAL MACHINE CONCURRENT PASCAL MACHINE. STORE ALLOCATION Per Brinch Hansen Information Science California Institute of Technology Pasadena, California 91125 October 1975 Summary This describes the allocation of core store among the classes, monitors, and processes of a Concurrent Pascal program running on a PDP 11/45 computer. Key Words and Phrasesl Concurrent Pascal implementation, Virtual machine, Store allocation, PDP 11/45 computer. CR Categories I 4.2, 6.2 The development of Concurrent Pascal has been supported by the National Science Foundation under grant number DCR74-17331. Copyright ~ 1975 Per Brlnch Hansen CONCURRENT PASCAL MACHINE Per Brinch Hansen InforMation Science California Institute of Technology Pasadena. California 91125 October 1975 This is a description of the virtual machine that executes Concurrent Pascal programs on the PDP 11/45 computer. It contains four papers entitled. 1. Concurrent Pascal machine. store allocation 2. Concurrent Pascal machinel Code interpretation 3. Concurrent Pascal machine: Kernel 4. The Concurrent Pascal compiler The development of Concurrent Pascal has been supported by the National Science foundation under grant number DCR74-17331. Copyright ~ 1975 Per Brinch Hansen CONTENTS 1 • INTRODUCTION 1 2. CORE STORE 1 3. VIRTUAL STORE 3 4. DATA SEGMENTS 5 5. PERMANENT VARIABLES 6 6, TEMPORARY VARIABLES 7 7. COMPRomISES 8 REfERENCES 10 1. INTRODUCTION Concurrent Pascal is a programming language for structured programming of computer operating systems [1, 2]. The Concurrent Pascal compiler generates code for a virtual machine that can be simulated by microprogram or machine code on different computers [3] . This paper describes the allocation of core store among the processes of a Concurrent Pascal program running on a PDP 11/45 computer.
    [Show full text]
  • Monitors and Concurrent Pascal: a Personal History
    Monitors and Concurrent Pascal: A Personal History (1993) This is a personal history of the early development of the monitor concept and its implementation in the programming language Concurrent Pascal. The paper explains how monitors evolved from the ideas of Dahl, Dijkstra, Hoare, and the author (1971–73). At Caltech the author and his students developed and implemented Concurrent Pascal and used it to write several model op- erating systems (1974–75). A portable implementation of Concurrent Pascal was widely distributed and used for system design (1976–90). The monitor paradigm was also disseminated in survey papers and text books. The au- thor ends the story by expressing his own mixed feelings about monitors and Concurrent Pascal. 1 A Programming Revolution In the 1970s new programming languages were developed to express asyn- chronous, concurrent processes. These languages support the now familiar paradigms for process communication known as monitors, remote procedure calls, and synchronous communication. The most inuential early idea was the monitor concept and its implementation in the programming language Concurrent Pascal. This is a personal history of how monitors and Concurrent Pascal were invented. I have tried to write the history of an idea—how it arose and spread through the scientic community. I have also described the struggles of the creative process, how you grope your way through obscurities and blind alleys until you nd an elegant way of expressing an idea. P. Brinch Hansen, Monitors and Concurrent Pascal: A personal history. 2nd ACM Conference on the History of Programming Languages, Cambridge, MA, April 1993. In SIGPLAN Notices 28, 3 (March 1993), 1–35.
    [Show full text]
  • The Invention of Concurrent Programming
    THE INVENTION OF CONCURRENT PROGRAMMING PER BRINCH HANSEN (2001) The author selects classic papers written by the computer scientists who made the major breakthroughs in concurrent programming. These papers cover the pioneering era of the eld from the semaphores of the mid 1960s to the remote procedure calls of the late 1970s. The author summarizes the classic papers and puts them in historical perspective. A PROGRAMMING REVOLUTION This is the story of one of the major revolutions in computer programming: the invention of concurrent programming. Tom Kilburn and David Howarth pioneered the use of interrupts to simu- late concurrent execution of several programs on the Atlas computer (Kilburn 1961). This programming technique became known as multiprogramming. The early multiprogramming systems were programmed in assembly lan- guage without any conceptual foundation. The slightest programming mis- take could make these systems behave in a completely erratic manner that made program testing nearly impossible. By the end of the 1960s multiprogrammed operating systems had become so huge and unreliable that their designers spoke openly of a software crisis (Naur 1969). As J. M. Havender (1968) recalled: P. Brinch Hansen, The invention of concurrent programming. In The Origin of Concur- rent Programming: From Semaphores to Remote Procedure Calls, P. Brinch Hansen, Ed., Copyright c 2001, Springer-Verlag, New York. 1 2 PER BRINCH HANSEN The original multitasking concept of the [IBM OS/360] envisioned rel- atively unrestrained competion for resources to perform a number of tasks concurrently ... But as the system evolved many instances of task deadlock were uncovered. Elliott Organick (1973) pointed out that the termination of a task in the Burroughs B6700 system might cause its ospring tasks to lose their stack space! In the mid 1960s computer scientists took the rst steps towards a deeper understanding of concurrent programming.
    [Show full text]