Openvms Programming Concepts Manual, Vol. 2

Total Page:16

File Type:pdf, Size:1020Kb

Openvms Programming Concepts Manual, Vol. 2 OpenVMS Programming Concepts Manual, Volume II Order Number: AA–PV67E–TK April 2001 This manual is volume two of two volumes. It describes the features that the OpenVMS operating system provides to programmers. Revision/Update Information: This manual supersedes the OpenVMS Programming Concepts Manual, Version 7.2. Software Version: OpenVMS Alpha Version 7.3 OpenVMS VAX Version 7.3 Compaq Computer Corporation Houston, Texas © 2001 Compaq Computer Corporation Compaq, AlphaServer, VAX, VMS, and the Compaq logo Registered in U.S. Patent and Trademark Office. Alpha, OpenVMS, PATHWORKS, DECnet, and DEC are trademarks of Compaq Information Technologies Group, L.P. in the United States and other countries. UNIX and X/Open are trademarks of The Open Group in the United States and other countries. All other product names mentioned herein may be trademarks of their respective companies. Confidential computer software. Valid license from Compaq required for possession, use, or copying. Consistent with FAR 12.211 and 12.212, Commercial Computer Software, Computer Software Documentation, and Technical Data for Commercial Items are licensed to the U.S. Government under vendor’s standard commercial license. Compaq shall not be liable for technical or editorial errors or omissions contained herein. The information in this document is provided "as is" without warranty of any kind and is subject to change without notice. The warranties for Compaq products are set forth in the express limited warranty statements accompanying such products. Nothing herein should be construed as constituting an additional warranty. Printed in the U.S.A. ZK6641 The Compaq OpenVMS documentation set is available on CD–ROM. Contents Preface ............................................................ xvii Part I OpenVMS Programming Interfaces: Calling a System Routine 17 Call Format to OpenVMS Routines 17.1 Overview . ................................................ 17–1 17.2 Format Heading ............................................. 17–2 17.2.1 Procedure Call Format ..................................... 17–3 17.2.2 JSB Call Format (VAX only) ................................ 17–4 17.3 Returns Heading ............................................ 17–4 17.3.1 Condition Values Returned in a Register ....................... 17–4 17.3.2 Other Returned Values .................................... 17–5 17.3.3 Condition Values Signaled . ................................ 17–6 17.4 Arguments Heading . ........................................ 17–6 17.4.1 OpenVMS Usage Entry .................................... 17–6 17.4.2 Type Entry .............................................. 17–7 17.4.3 Access Entry ............................................ 17–8 17.4.4 Mechanism Entry . ........................................ 17–9 17.4.5 Explanatory Text . ........................................ 17–10 17.5 Condition Values Returned Heading ............................. 17–11 17.5.1 Condition Values Returned . ................................ 17–12 17.5.2 Condition Values Returned in an I/O Status Block ............... 17–12 17.5.3 Condition Values Returned in a Mailbox ....................... 17–12 17.5.4 Condition Values Signaled . ................................ 17–12 18 Basic Calling Standard Conventions 18.1 Hardware Registers . ........................................ 18–1 18.1.1 Register Usage for OpenVMS VAX ............................ 18–1 18.1.2 Register Usage for OpenVMS Alpha . ....................... 18–2 18.1.2.1 Integer Registers ...................................... 18–2 18.1.2.2 Floating-Point Registers . ................................ 18–3 18.2 Stack Usage for Procedures .................................... 18–3 18.2.1 Stack Procedure Usage for VAX .............................. 18–3 18.2.1.1 Calling Sequence ...................................... 18–5 18.2.1.2 Call Frames on Return . ................................ 18–5 18.2.2 Stack Procedure Usage for Alpha ............................. 18–5 18.2.2.1 Fixed-Size Stack Frame . ................................ 18–5 18.2.2.2 Variable-Size Stack Frame ............................... 18–6 18.3 Procedure Representation ..................................... 18–8 18.4 Argument List .............................................. 18–9 18.5 Argument Passing Mechanisms . ................................ 18–11 18.5.1 Passing Arguments by Value ................................ 18–13 iii 18.5.2 Passing Arguments by Reference . ............................ 18–14 18.5.3 Passing Arguments by Descriptor ............................ 18–16 18.6 Passing Scalars as Arguments .................................. 18–18 18.7 Passing Arrays as Arguments .................................. 18–18 18.8 Passing Strings as Arguments .................................. 18–18 18.9 Combinations of Descriptor Class and Data Type ................... 18–19 18.10 Return of the Function Value ................................... 18–22 18.11 Return of the Condition Value .................................. 18–22 19 Calling Run-Time Library Routines 19.1 Overview .................................................. 19–1 19.2 Call Instructions ............................................ 19–3 19.2.1 Facility Prefix and Routine Name ............................ 19–3 19.2.2 The RTL Call Entry . .................................... 19–4 19.2.2.1 JSB Call Entries (VAX Only) . ............................ 19–6 19.2.3 Returns from an RTL Routine . ............................ 19–6 19.2.3.1 Facility Return Status and Condition Value Symbols .......... 19–6 19.3 Calling a Library Procedure in VAX MACRO (VAX Only) . ........... 19–7 19.3.1 VAX MACRO Calling Sequence . ............................ 19–7 19.3.2 VAX MACRO CALLS Instruction Example . ................... 19–8 19.3.3 VAX MACRO CALLG Instruction Example . ................... 19–8 19.3.4 VAX MACRO JSB Entry Points . ............................ 19–9 19.3.5 Return Status ........................................... 19–10 19.3.6 Function Return Values in VAX MACRO ....................... 19–11 19.4 Calling a Library Routine in BLISS . ............................ 19–11 19.4.1 BLISS Calling Sequence ................................... 19–12 19.4.2 Accessing a Return Status in BLISS .......................... 19–12 19.4.3 Calling JSB Entry Points from BLISS ......................... 19–13 20 Calling System Services 20.1 Overview .................................................. 20–1 20.2 Preserving System Integrity ................................... 20–2 20.2.1 User Privileges ........................................... 20–2 20.2.2 Resource Quotas ......................................... 20–2 20.2.3 Access Modes ............................................ 20–2 20.3 System Service Call Entry . .................................... 20–3 20.4 System Service Completion .................................... 20–5 20.4.1 Asynchronous and Synchronous System Services ................. 20–5 20.4.2 System Service Resource Wait Mode .......................... 20–6 20.4.3 Condition Values Returned from System Services ................ 20–7 20.4.4 Testing the Condition Value................................. 20–8 20.4.5 Special Condition Values Using Symbolic Codes ................. 20–8 20.4.6 Testing the Return Condition Value for Assembly Languages ....... 20–9 20.4.7 System Messages Generated by Condition Values ................ 20–10 20.5 Program Examples with System Service Calls . ................... 20–11 iv 21 STARLET Structures and Definitions for C Programmers 21.1 SYS$STARLET_C.TLB Equivalency to STARLETSD.TLB ............. 21–1 21.2 NEW STARLET Definitions for C ............................... 21–2 Part II I/O, System and Programming Routines 22 Run-Time Library Input/Output Operations 22.1 Choosing I/O Techniques ...................................... 22–1 22.2 Using SYS$INPUT and SYS$OUTPUT ........................... 22–3 22.2.1 Default Input and Output Devices............................ 22–3 22.2.2 Reading and Writing to Alternate Devices and External Files ....... 22–4 22.3 Working with Simple User I/O . ................................ 22–4 22.3.1 Default Devices for Simple I/O ............................... 22–4 22.3.2 Getting a Line of Input .................................... 22–4 22.3.3 Getting Several Lines of Input ............................... 22–5 22.3.4 Writing Simple Output ..................................... 22–6 22.4 Working with Complex User I/O ................................ 22–7 22.4.1 Compaq DECwindows Motif . ................................ 22–7 22.4.1.1 DECwindows Server Height or Width Exceeding 32767 (VAX Only) ............................................... 22–8 22.4.1.2 SET DISPLAY Used to Create WSA Pseudo Workstation Devices .............................................. 22–8 22.4.2 SMG$ Run-Time Routines . ................................ 22–9 22.4.3 Pasteboards ............................................. 22–10 22.4.3.1 Erasing a Pasteboard . ................................ 22–11 22.4.3.2 Deleting a Pasteboard . ................................ 22–11 22.4.3.3 Setting Screen Dimensions and Background Color ............ 22–11 22.4.4 Virtual Displays . ........................................ 22–12 22.4.4.1 Creating a Virtual Display ............................... 22–12 22.4.4.2 Pasting Virtual Displays ................................ 22–13 22.4.4.3 Rearranging Virtual Displays ............................ 22–15 22.4.4.4 Removing Virtual Displays ..............................
Recommended publications
  • Demystifying the Real-Time Linux Scheduling Latency
    Demystifying the Real-Time Linux Scheduling Latency Daniel Bristot de Oliveira Red Hat, Italy [email protected] Daniel Casini Scuola Superiore Sant’Anna, Italy [email protected] Rômulo Silva de Oliveira Universidade Federal de Santa Catarina, Brazil [email protected] Tommaso Cucinotta Scuola Superiore Sant’Anna, Italy [email protected] Abstract Linux has become a viable operating system for many real-time workloads. However, the black-box approach adopted by cyclictest, the tool used to evaluate the main real-time metric of the kernel, the scheduling latency, along with the absence of a theoretically-sound description of the in-kernel behavior, sheds some doubts about Linux meriting the real-time adjective. Aiming at clarifying the PREEMPT_RT Linux scheduling latency, this paper leverages the Thread Synchronization Model of Linux to derive a set of properties and rules defining the Linux kernel behavior from a scheduling perspective. These rules are then leveraged to derive a sound bound to the scheduling latency, considering all the sources of delays occurring in all possible sequences of synchronization events in the kernel. This paper also presents a tracing method, efficient in time and memory overheads, to observe the kernel events needed to define the variables used in the analysis. This results in an easy-to-use tool for deriving reliable scheduling latency bounds that can be used in practice. Finally, an experimental analysis compares the cyclictest and the proposed tool, showing that the proposed method can find sound bounds faster with acceptable overheads. 2012 ACM Subject Classification Computer systems organization → Real-time operating systems Keywords and phrases Real-time operating systems, Linux kernel, PREEMPT_RT, Scheduling latency Digital Object Identifier 10.4230/LIPIcs.ECRTS.2020.9 Supplementary Material ECRTS 2020 Artifact Evaluation approved artifact available at https://doi.org/10.4230/DARTS.6.1.3.
    [Show full text]
  • A Case Against the GOT' O William a . Wulf, Carnegie-Mellon Universit Y
    A Case Against the GOT' O William A . Wulf, Carnegie-Mellon Universit y ABSTRAC T suggestion to ban the goto appears to have been a part of the computing folklore for several years , It has been proposed, by E . W . Dijkstra and others , to this author's knowledge the suggestion wa s that the goto statement in programming language i s first made in print by Professor E . W . Dijkstra i n a principal culprit in programs which are diffi- a letter to the editor of the _Communications o f cult to understand, modify, and debug . More cor- the ACM in 1968 (1) . rectly, the argument is that it is possible t o In this paper we shall examine the rational e use the goto to synthesize program structures wit h for the elimination of the Soto in programin g these undesirable properties . Not all uses of th e languages, and some of the theoretical and practi- goto are to be considered harmful ; however, it i s cal implications of its (total) elimination . further argued that the "good" uses of the got o fall into one of a small number of specific case s RATIONAL E which may be handled by specific language con- structs . This paper summarizes the arguments i n At one level, the rationale for eliminatin g favor of eliminating the goto statement and som e the poto has already been given in the introduc- of the theoretical and practical implications o f tion . Namely, it is possible to use the goto in a the proposal . manner which obscures the logical structure of a program to a point where it becomes virtually im- KEY WORDS AND PHRASES : programming, programmin g possible to understand (1,3,4), It is not claimed languages, goto-less programming, structured pro- that every use of the goto obscures the logica l grammin g structure of a program ; it is only claimed that i t CR CATEGORIES : 4 .2, 4 .22, 5 .2 4 is possible to use the Soto to fabricate a "rat' s nest" of control flow which has the undesirabl e INTRODUCTION properties mentioned above .
    [Show full text]
  • Context Switch in Linux – OS Course Memory Layout – General Picture
    ContextContext switchswitch inin LinuxLinux ©Gabriel Kliot, Technion 1 Context switch in Linux – OS course Memory layout – general picture Stack Stack Stack Process X user memory Process Y user memory Process Z user memory Stack Stack Stack tss->esp0 TSS of CPU i task_struct task_struct task_struct Process X kernel Process Y kernel Process Z kernel stack stack and task_struct stack and task_struct and task_struct Kernel memory ©Gabriel Kliot, Technion 2 Context switch in Linux – OS course #1 – kernel stack after any system call, before context switch prev ss User Stack esp eflags cs … User Code eip TSS … orig_eax … tss->esp0 es Schedule() function frame esp ds eax Saved on the kernel stack during ebp a transition to task_struct kernel mode by a edi jump to interrupt and by SAVE_ALL esi macro edx thread.esp0 ecx ebx ©Gabriel Kliot, Technion 3 Context switch in Linux – OS course #2 – stack of prev before switch_to macro in schedule() func prev … Schedule() saved EAX, ECX, EDX Arguments to contex_switch() Return address to schedule() TSS Old (schedule’s()) EBP … tss->esp0 esp task_struct thread.eip thread.esp thread.esp0 ©Gabriel Kliot, Technion 4 Context switch in Linux – OS course #3 – switch_to: save esi, edi, ebp on the stack of prev prev … Schedule() saved EAX, ECX, EDX Arguments to contex_switch() Return address to schedule() TSS Old (schedule’s()) EBP … tss->esp0 ESI EDI EBP esp task_struct thread.eip thread.esp thread.esp0 ©Gabriel Kliot, Technion 5 Context switch in Linux – OS course #4 – switch_to: save esp in prev->thread.esp
    [Show full text]
  • What Is an Operating System III 2.1 Compnents II an Operating System
    Page 1 of 6 What is an Operating System III 2.1 Compnents II An operating system (OS) is software that manages computer hardware and software resources and provides common services for computer programs. The operating system is an essential component of the system software in a computer system. Application programs usually require an operating system to function. Memory management Among other things, a multiprogramming operating system kernel must be responsible for managing all system memory which is currently in use by programs. This ensures that a program does not interfere with memory already in use by another program. Since programs time share, each program must have independent access to memory. Cooperative memory management, used by many early operating systems, assumes that all programs make voluntary use of the kernel's memory manager, and do not exceed their allocated memory. This system of memory management is almost never seen any more, since programs often contain bugs which can cause them to exceed their allocated memory. If a program fails, it may cause memory used by one or more other programs to be affected or overwritten. Malicious programs or viruses may purposefully alter another program's memory, or may affect the operation of the operating system itself. With cooperative memory management, it takes only one misbehaved program to crash the system. Memory protection enables the kernel to limit a process' access to the computer's memory. Various methods of memory protection exist, including memory segmentation and paging. All methods require some level of hardware support (such as the 80286 MMU), which doesn't exist in all computers.
    [Show full text]
  • Does the Seventh Amendment Compel Fact-‐Finding Before
    HAVE WE GONE TOO FAR: DOES THE SEVENTH AMENDMENT COMPEL FACT-FINDING BEFORE REACHING A DECISION ON PATENT- ELIGIBLE SUBJECT MATTER? JESSE D.H. SNYDER* I. INTRODUCTION Something seems amiss. Settled notions of what can be Patenta- ble are changing.1 To complicate matters further, so far, neither the SuPreme Court nor the Federal Circuit has addressed whether the in- quiry into Patent-eliGible subject matter is factual, legal, or both. And while the scoPe of Patentable subject matter is understandably con- strictinG, the end result may not be coterminous with the SuPreme Court’s Guidance in Alice Corporation Party Limited v. CLS Bank.2 Under Alice, which built uPon the Patent-eliGible subject matter frameworks of Prior cases, for those Patent claims that are directed to a Patent- ineligible concePt (e.G., abstract ideas), Patentability turns on whether an “inventive concePt” exists to “‘transform the nature of the claim’ into a Patent-eliGible aPPlication.”3 The lower courts have seized on the Alice holding, inhuming Pa- tents whose subject-matter failures render them invalid as a matter of law. But what if the means adoPted by the lower courts to reach that leGal conclusion also sidestePs certain constitutional guarantees? What if they are simPlifyinG the analysis at a facial level when a more granular investigation is required? What if, at the motion to dismiss staGe, fact-findinG is still required to determine whether a Patent is * 2015-2016 term law clerk to the Honorable Jimmie V. Reyna of the U.S. Court of APPeals for the Federal Circuit. I earned my J.D., summa cum laude, from Texas Wesleyan School of Law and my B.S.
    [Show full text]
  • Hiding Process Memory Via Anti-Forensic Techniques
    DIGITAL FORENSIC RESEARCH CONFERENCE Hiding Process Memory via Anti-Forensic Techniques By: Frank Block (Friedrich-Alexander Universität Erlangen-Nürnberg (FAU) and ERNW Research GmbH) and Ralph Palutke (Friedrich-Alexander Universität Erlangen-Nürnberg) From the proceedings of The Digital Forensic Research Conference DFRWS USA 2020 July 20 - 24, 2020 DFRWS is dedicated to the sharing of knowledge and ideas about digital forensics research. Ever since it organized the first open workshop devoted to digital forensics in 2001, DFRWS continues to bring academics and practitioners together in an informal environment. As a non-profit, volunteer organization, DFRWS sponsors technical working groups, annual conferences and challenges to help drive the direction of research and development. https://dfrws.org Forensic Science International: Digital Investigation 33 (2020) 301012 Contents lists available at ScienceDirect Forensic Science International: Digital Investigation journal homepage: www.elsevier.com/locate/fsidi DFRWS 2020 USA d Proceedings of the Twentieth Annual DFRWS USA Hiding Process Memory Via Anti-Forensic Techniques Ralph Palutke a, **, 1, Frank Block a, b, *, 1, Patrick Reichenberger a, Dominik Stripeika a a Friedrich-Alexander Universitat€ Erlangen-Nürnberg (FAU), Germany b ERNW Research GmbH, Heidelberg, Germany article info abstract Article history: Nowadays, security practitioners typically use memory acquisition or live forensics to detect and analyze sophisticated malware samples. Subsequently, malware authors began to incorporate anti-forensic techniques that subvert the analysis process by hiding malicious memory areas. Those techniques Keywords: typically modify characteristics, such as access permissions, or place malicious data near legitimate one, Memory subversion in order to prevent the memory from being identified by analysis tools while still remaining accessible.
    [Show full text]
  • Context Switch Overheads for Linux on ARM Platforms
    Context Switch Overheads for Linux on ARM Platforms Francis David [email protected] Jeffrey Carlyle [email protected] Roy Campbell [email protected] http://choices.cs.uiuc.edu Outline What is a Context Switch? ◦ Overheads ◦ Context switching in Linux Interrupt Handling Overheads ARM Experimentation Platform Context Switching ◦ Experiment Setup ◦ Results Interrupt Handling ◦ Experiment Setup ◦ Results What is a Context Switch? Storing current processor state and restoring another Mechanism used for multi-tasking User-Kernel transition is only a processor mode switch and is not a context switch Sources of Overhead Time spent in saving and restoring processor state Pollution of processor caches Switching between different processes ◦ Virtual memory maps need to be switched ◦ Synchronization of memory caches Paging Context Switching in Linux Context switch can be implemented in userspace or in kernelspace New 2.6 kernels use Native POSIX Threading Library (NPTL) NPTL uses one-to-one mapping of userspace threads to kernel threads Our experiments use kernel 2.6.20 Outline What is a Context Switch? ◦ Overheads ◦ Context switching in Linux Interrupt Handling Overheads ARM Experimentation Platform Context Switching ◦ Experiment Setup ◦ Results Interrupt Handling ◦ Experiment Setup ◦ Results Interrupt Handling Interruption of normal program flow Virtual memory maps not switched Also causes overheads ◦ Save and restore of processor state ◦ Perturbation of processor caches Outline What is a Context Switch? ◦ Overheads ◦ Context switching
    [Show full text]
  • Lecture 4: September 13 4.1 Process State
    CMPSCI 377 Operating Systems Fall 2012 Lecture 4: September 13 Lecturer: Prashant Shenoy TA: Sean Barker & Demetre Lavigne 4.1 Process State 4.1.1 Process A process is a dynamic instance of a computer program that is being sequentially executed by a computer system that has the ability to run several computer programs concurrently. A computer program itself is just a passive collection of instructions, while a process is the actual execution of those instructions. Several processes may be associated with the same program; for example, opening up several windows of the same program typically means more than one process is being executed. The state of a process consists of - code for the running program (text segment), its static data, its heap and the heap pointer (HP) where dynamic data is kept, program counter (PC), stack and the stack pointer (SP), value of CPU registers, set of OS resources in use (list of open files etc.), and the current process execution state (new, ready, running etc.). Some state may be stored in registers, such as the program counter. 4.1.2 Process Execution States Processes go through various process states which determine how the process is handled by the operating system kernel. The specific implementations of these states vary in different operating systems, and the names of these states are not standardised, but the general high-level functionality is the same. When a process is first started/created, it is in new state. It needs to wait for the process scheduler (of the operating system) to set its status to "new" and load it into main memory from secondary storage device (such as a hard disk or a CD-ROM).
    [Show full text]
  • Shri Vaishnav Vidyapeeth Vishwavidyalaya, Indore
    SHRI VAISHNAV VIDYAPEETH VISHWAVIDYALAYA, INDORE Memories are special moments that tell our story SMRITI मृत ALUMNI DIRECTORY 2020 The rising flame epitomises leadership through enlightenment The bright orange colour represents brilliance and colour blue reflects serenity and infinity तमसो मा योतगमय Lead me from darkness to light ABOUT UNIVERSITY Shri Vaishnav Vidyapeeth Vishwavidyalaya is a state private university established under Madhya Pradesh Niji Vishwavidyalaya (Sthapana Avam Sanchalan) Adhiniyam in 2015 at Indore, MP(India). The University has been established with a vision to be a leader in shaping better future for mankind through quality education, training and research. It shall pursue the mission to make a difference in sustaining the growth of global societies by developing socially responsible citizens. Value based education being at the helm, the University offers degree and diploma programs through 13 constituent institutes and schools, besides Faculty of Doctoral Studies and Research. Some of the objectives of the University are as under: • To provide teaching and training in higher education and make provision for research as well as advancement and dissemination of knowledge. • To ensure world class quality in its offerings and create higher levels of intellectual abilities. • To create centres of excellence for research and development for sharing knowledge and its applications. Shri Vaishnav Vidyapeeth Vishwavidyalaya at Indore is a multi-disciplinary university focusing on the needs of various segments of the society. FROM THE DESK OF CHANCELLOR “To emerge as a winner, learn to balance relaxation and action” Shri Vaishnav Vidyapeeth Vishwavidyalaya was conceived as an institution of higher learning for imparting quality education to the young and experienced learners alike.
    [Show full text]
  • IBM Power Systems Virtualization Operation Management for SAP Applications
    Front cover IBM Power Systems Virtualization Operation Management for SAP Applications Dino Quintero Enrico Joedecke Katharina Probst Andreas Schauberer Redpaper IBM Redbooks IBM Power Systems Virtualization Operation Management for SAP Applications March 2020 REDP-5579-00 Note: Before using this information and the product it supports, read the information in “Notices” on page v. First Edition (March 2020) This edition applies to the following products: Red Hat Enterprise Linux 7.6 Red Hat Virtualization 4.2 SUSE Linux SLES 12 SP3 HMC V9 R1.920.0 Novalink 1.0.0.10 ipmitool V1.8.18 © Copyright International Business Machines Corporation 2020. All rights reserved. Note to U.S. Government Users Restricted Rights -- Use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM Corp. Contents Notices . .v Trademarks . vi Preface . vii Authors. vii Now you can become a published author, too! . viii Comments welcome. viii Stay connected to IBM Redbooks . ix Chapter 1. Introduction. 1 1.1 Preface . 2 Chapter 2. Server virtualization . 3 2.1 Introduction . 4 2.2 Server and hypervisor options . 4 2.2.1 Power Systems models that support PowerVM versus KVM . 4 2.2.2 Overview of POWER8 and POWER9 processor-based hardware models. 4 2.2.3 Comparison of PowerVM and KVM / RHV . 7 2.3 Hypervisors . 8 2.3.1 Introducing IBM PowerVM . 8 2.3.2 Kernel-based virtual machine introduction . 15 2.3.3 Resource overcommitment . 16 2.3.4 Red Hat Virtualization . 17 Chapter 3. IBM PowerVM management and operations . 19 3.1 Shared processor logical partitions . 20 3.1.1 Configuring a shared processor LPAR .
    [Show full text]
  • Lecture 6: September 20 6.1 Threads
    CMPSCI 377 Operating Systems Fall 2012 Lecture 6: September 20 Lecturer: Prashant Shenoy TA: Sean Barker & Demetre Lavigne 6.1 Threads A thread is a sequential execution stream within a process. This means that a single process may be broken up into multiple threads. Each thread has its own Program Counter, registers, and stack, but they all share the same address space within the process. The primary benefit of such an approach is that a process can be split up into many threads of control, allowing for concurrency since some parts of the process to make progress while others are busy. Sharing an address space within a process allows for simpler coordination than message passing or shared memory. Threads can also be used to modularize a process { a process like a web browser may create one thread for each browsing tab, or create threads to run external plugins. 6.1.1 Processes vs threads One might argue that in general processes are more flexible than threads. For example, processes are controlled independently by the OS, meaning that if one crashes it will not affect other processes. However, processes require explicit communication using either message passing or shared memory which may add overhead since it requires support from the OS kernel. Using threads within a process allows them all to share the same address space, simplifying communication between threads. However, threads have their own problems: because they communicate through shared memory they must run on same machine and care must be taken to create thread-safe code that functions correctly despite multiple threads acting on the same set of shared data.
    [Show full text]
  • Mitigating the Performance-Efficiency Tradeoff in Resilient Memory Disaggregation
    Mitigating the Performance-Efficiency Tradeoff in Resilient Memory Disaggregation Youngmoon Lee∗, Hasan Al Maruf∗, Mosharaf Chowdhury∗, Asaf Cidonx, Kang G. Shin∗ University of Michigan∗ Columbia Universityx ABSTRACT State-of-the-art solutions take three primary approaches: (i) local We present the design and implementation of a low-latency, low- disk backup [36, 66], (ii) remote in-memory replication [31, 50], and overhead, and highly available resilient disaggregated cluster mem- (iii) remote in-memory erasure coding [61, 64, 70, 73] and compres- ory. Our proposed framework can access erasure-coded remote sion [45]. Unfortunately, they suffer from some combinations of memory within a single-digit `s read/write latency, significantly the following problems. improving the performance-efficiency tradeoff over the state-of- High latency: The first approach has no additional memory the-art – it performs similar to in-memory replication with 1.6× overhead, but the access latency is intolerably high in the presence lower memory overhead. We also propose a novel coding group of any of the aforementioned failure scenarios. The systems that placement algorithm for erasure-coded data, that provides load bal- take the third approach do not meet the single-digit `s latency ancing while reducing the probability of data loss under correlated requirement of disaggregated cluster memory even when paired failures by an order of magnitude. with RDMA (Figure 1). High cost: While the second approach has low latency, it dou- bles memory consumption as well as network bandwidth require- ments. The first and second approaches represent the two extreme 1 INTRODUCTION points in the performance-vs-efficiency tradeoff space for resilient To address the increasing memory pressure in datacenters, two cluster memory (Figure 1).
    [Show full text]