Introduction to Operating Systems

Total Page:16

File Type:pdf, Size:1020Kb

Introduction to Operating Systems Introduction to Operating Systems Jakub Olejnik These slides are a much inexact translation of the material available at http://wazniak.mimuw.edu.pl. The authors of the original slides were kind enough to share their files with me thus help me in translation. Literature Lecture • A. Silberschatz „Operating Systems Concepts” • A. Tanenbaum „Modern Operating Systems” • webpage (in polish) – the OS section http://wazniak.mimuw.edu.pl Laboratory classes • C. Newham, B. Rosenblatt „Learning the bash shell” • Unix/Linux manual pages Basic Concepts of Operating Systems Lecture 1 What’s in this lecture? • Definition of an Operating System, its place and its role in computer system. • What kind of Operating Systems are there? • Operating Systems Concepts – how do the OS’s work? What is an Operating System? A definition An Operating System is a set of computer programs which operate the computer’s devices directly and whose task is to present to a user an unified environment for writing and running programs in an convenient and efficient way. thus… An OS is a proxy between the user programs and the actual hardware. User and his programs virtual machine API Operating System drivers Actual Hardware Structure of an Operating System There are basicaly three types of OS’ components i.e.: graphical or text-based shells The operating system system programs: `CD’ commands, adm consoles Kernel modules hardware The Tasks of Operating System • Defining the user interface (shell) • Providing a file system • Providing environment to run user programs – mechanisms of actual code execution, safety. – mechanisms of process synchronization, communication • Managing the I/O devices • Providing procedures for handling basic errors Resource Management Resource Management Resource – any software or hardware component essential to processing • Resource allocation – assigning resources to processes • Resource allocation planning – choosing optimal allocation strategy • Access authorization and resource protection – restricting resource access in accordance with permissions granted • Resource recycling – managing reusable resources • Accounting – maintaining access register Resources managed by Operating System • Processor – each program gets a chance to be executed • Memory – allocation of memory space to processes – memory protection and address transformation • I/O devices – providing access to keyboards, mice, mass storage devices, printers, scanners, plotters etc. • Information (file system) – organizing information on disks, providing access – protection and access authorization Basic classification of Operating Systems OS’s classified with respect to processing • Interactive Systems (on-line systems) – users directly interact with operating system – task execution starts immediately after submission • Common OS’s serve as examples • Batch Systems (off-line systems) – task execution can be delayed – users can not interfere with task execution after their task has been submitted • Not many examples of pure off-line systems • Partial examples include: mainframes, job scheduling, print spooling, bulk transaction processing eg. ELIXIR Batch System – a historical example Early batch system – Programmers bring cards to 1401 – Operator reads cards onto tape – puts tape on 7094 which does computing – puts tape on 1401 which prints output OS’s classified with respect to number of programs executed • Single-tasking systems – it is not possible to start a new user program until previous task has finished – Examples: the well-known MS-DOS and many earlier OS’s • Multi-tasking/Multiprogramming systems – Multiple can be executed `at the same time’ – OS’s provides mechanism of context switching to time- share the CPU – Basic strategies: pre-emptive, co-operative multitasking, real-time systems OS’s classified with respect to number of users supported • Single-user systems – all of computer’s resources are accessed by only one user thus mechanisms of protection and authorization are not necessary – Examples: MS-DOS, AmigaOS, Win95/98, WinXP home ed. – in a way. • Multi-user systems – Many users can access computer’s resources thus mechanisms of protection and authorization are vital – Examples: UNIX, Linux, Windows server families Other types of OS’s • Real-time operating systems • Distributed (network) systems • Server, Mainframe systems • PC systems • Multiprocessor systems • Embedded systems • Smart card systems How does an Operating System work? Von Neumann architecture • Instruction codes as well as operands (args to instr.) are stored in the memory • Instructions are executed in the order in which they are located in memory • Memory location of next command is stored in PC/IP • The order of instruction execution can be changed by flow control instructions (jumps, cond. jumps by PC:=PC + label) • After instruction execution the instruction *(PC) is loaded, decoded and PC:=PC + length(*PC) Instruction cycle in detail – 1 • Instruction cycle (FDX-cycle) – the sequence of actions that the processor (CPU) performs to execute each machine code instruction in a program. • Instruction cycle consists of phases. • Typical phases of FDX-cycle are – instruction loading phase – memory is read – operand loading phase – memory is read – instruction execution phase – operand storing phase – memory is written Instruction cycle in detail – 2 1. Fetch the instruction – PC is presented on address bus, instruction code is read from data bus to CIR 2. Instruction is decoded by decoding unit 3. Fetch operands – Operand's effective address is fetched from memory (optionally), the operand itself is fetched 4. Execution 5. Store results – Operand is stored in the memory 6. Check for interrupts if an interrupt has been raised its number (source) is identified and the context is switched to appropriate interrupt handle routine if no interrupt has been raised go to 1. Modern CPUs Nowadays, processors implement mechanisms to speed up the instruction cycle: • Pipelines – when one instruction is being decoded why not fetch the next one • Superscalar execution – why not have more execution units • Multithreading – multiple threads parallelism • Multi-core CPUs – multithreading without sharing computing unit, CPU cashe, TLB Interrupts An interrupt is a state of computer system in reaction to an asynchronous event (an event which happens independently of processing and its precise moment of occurrence cannot be predicted) . It consists in the processor automatically saving its state of execution (context switch), and starting execution of an interrupt handler. Sources of Interrupts • External interrupts – raised by external devices • Software interrupts – raised by executed program with special instruction (interrupt call) • Diagnostic interrupts (exception handling) – raised to handle an exceptional state of processing: – traps – software errors e.g..: division by 0, memory protection violation, incorrect instruction; OS’s kernel removes the faulty process – hardware errors e.g..: segmentation fault Basics of OS’s operation • The OS’s kernel is called by the use interrupts, in this respect the kernel is a set of int. handlers and related procs. • Hardware memory protection • At least two CPU modes: (user mode, kernel mode) • A set of privileged instructions is available only in the kernel mode • the I/O instructions are privileged • a clock generates clock interrupts at constant intervals of time, so that periodically the kernel regains control Interrupts in the layer cake user software program interrupt external interrupt Operating System machine code diagnostic interrupt microarchitecture Hardware memory protection example • Executed program address makes reference to a specific location in memory • Memory chunks >= BASE which are available to the program are described by two fault registers: BASE and LIMIT < + LIMIT memory is accessed Scheduling by clock interrupts • Kernel regains control by clock interrupts • A clock interrupt (typically once per 10ms) runs a kernel procedure which chooses a program to be executed in turn • In single-tasking systems user programs (or viruses) sometimes modified the clock interrupt handler to emulate ‘double-tasking’. .
Recommended publications
  • A Programmable Microkernel for Real-Time Systems∗
    A Programmable Microkernel for Real-Time Systems∗ Christoph M. Kirsch Marco A.A. Sanvido Thomas A. Henzinger University of Salzburg VMWare Inc. EPFL and UC Berkeley [email protected] tah@epfl.ch ABSTRACT Categories and Subject Descriptors We present a new software system architecture for the im- D.4.7 [Operating Systems]: Organization and Design— plementation of hard real-time applications. The core of the Real-time systems and embedded systems system is a microkernel whose reactivity (interrupt handling as in synchronous reactive programs) and proactivity (task General Terms scheduling as in traditional RTOSs) are fully programma- Languages ble. The microkernel, which we implemented on a Strong- ARM processor, consists of two interacting domain-specific Keywords virtual machines, a reactive E (Embedded) machine and a proactive S (Scheduling) machine. The microkernel code (or Real Time, Operating System, Virtual Machine microcode) that runs on the microkernel is partitioned into E and S code. E code manages the interaction of the system 1. INTRODUCTION with the physical environment: the execution of E code is In [9], we advocated the E (Embedded) machine as a triggered by environment interrupts, which signal external portable target for compiling hard real-time code, and in- events such as the arrival of a message or sensor value, and it troduced, in [11], the S (Scheduling) machine as a universal releases application tasks to the S machine. S code manages target for generating schedules according to arbitrary and the interaction of the system with the processor: the exe- possibly non-trivial strategies such as nonpreemptive and cution of S code is triggered by hardware interrupts, which multiprocessor scheduling.
    [Show full text]
  • Interrupt Handling in Linux
    Department Informatik Technical Reports / ISSN 2191-5008 Valentin Rothberg Interrupt Handling in Linux Technical Report CS-2015-07 November 2015 Please cite as: Valentin Rothberg, “Interrupt Handling in Linux,” Friedrich-Alexander-Universitat¨ Erlangen-Nurnberg,¨ Dept. of Computer Science, Technical Reports, CS-2015-07, November 2015. Friedrich-Alexander-Universitat¨ Erlangen-Nurnberg¨ Department Informatik Martensstr. 3 · 91058 Erlangen · Germany www.cs.fau.de Interrupt Handling in Linux Valentin Rothberg Distributed Systems and Operating Systems Dept. of Computer Science, University of Erlangen, Germany [email protected] November 8, 2015 An interrupt is an event that alters the sequence of instructions executed by a processor and requires immediate attention. When the processor receives an interrupt signal, it may temporarily switch control to an inter- rupt service routine (ISR) and the suspended process (i.e., the previously running program) will be resumed as soon as the interrupt is being served. The generic term interrupt is oftentimes used synonymously for two terms, interrupts and exceptions [2]. An exception is a synchronous event that occurs when the processor detects an error condition while executing an instruction. Such an error condition may be a devision by zero, a page fault, a protection violation, etc. An interrupt, on the other hand, is an asynchronous event that occurs at random times during execution of a pro- gram in response to a signal from hardware. A proper and timely handling of interrupts is critical to the performance, but also to the security of a computer system. In general, interrupts can be emitted by hardware as well as by software. Software interrupts (e.g., via the INT n instruction of the x86 instruction set architecture (ISA) [5]) are means to change the execution context of a program to a more privileged interrupt context in order to enter the kernel and, in contrast to hardware interrupts, occur synchronously to the currently running program.
    [Show full text]
  • Interrupts and Exceptions CPU Modes and Address Spaces Dual-Mode Of
    CPU Modes and Address Spaces There are two processor (CPU) modes of operation: • Kernel (Supervisor) Mode and • User Mode The processor is in Kernel Mode when CPU mode bit in Status Interrupts and Exceptions register is set to zero. The processor enters Kernel Mode at power-up, or as result of an interrupt, exception, or error. The processor leaves Kernel Mode and enters User Mode when the CPU mode bit is set to one (by some instruction). Memory address space is divided in two ranges (simplified): • User address space – addresses in the range [0 – 7FFFFFFF16] Studying Assignment: A.7 • Kernel address space Reading Assignment: 3.1-3.2, A.10 – addresses in the range [8000000016 – FFFFFFFF16] g. babic Presentation B 28 g. babic 29 Dual-Mode of CPU Operation MIPS Privilege Instructions • CPU mode bit indicates the current CPU mode: 0 (=kernel) With CPU in User Mode, the program in execution has access or 1 (=user). only to the CPU and FPU registers, while when CPU operates • When an interrupt occurs, CPU hardware switches to the in Kernel Mode, the program has access to the full capabilities kernel mode. of processor including CP0 registers. • Switching to user mode (from kernel mode) done by setting CPU mode bit (by an instruction). Privileged instructions can not be executed when the Exception/Interrupt processor is in User mode, they can be executed only when the processor is in Kernel mode. kernel user Examples of MIPS privileged instructions: Set user mode • any instruction that accesses Kernel address space • all instructions that access any of CP0 registers, e.g.
    [Show full text]
  • Additional Functions in HW-RTOS Offering the Low Interrupt Latency
    HW-RTOS Real Time OS in Hardware Additional Functions in HW-RTOS Offering the Low Interrupt Latency In this white paper, we introduce two HW-RTOS functions that offer the lowest interrupt latency available and help create a more software-friendly environment. One of these is ISR implemented in hardware, which improves responsiveness when activating a task from an interrupt and eliminates the need for developing a handler in software. The other is a function allowing the use of non-OS managed interrupt handlers in a multitasking environment. This makes it easier to migrate from a non-RTOS environment to a multitasking one. R70WP0003EJ0100 September, 2018 2 / 8 Multitasking Environment with Lowest Interrupt Latency Offered by HW-RTOS 1. Executive Summary In this white paper, we introduce two functions special to HW-RTOS that improve interrupt performance. The first is the HW ISR function. Renesas stylized the ISR (Interrupt Service Routine) process and implemented it in hardware to create their HW ISR. With this function, the task corresponding to the interrupt signal can be activated directly and in real time. And, since the ISR is implemented in the hardware, application software engineers are relieved of the burden of developing a handler. The second is called Direct Interrupt Service. This function is equivalent to allowing a non-OS managed interrupt handler to invoke an API. This function %" "$# $""%!$ $ $""%!$!" enables synchronization and communication "$ "$ between the non-OS managed interrupt handler and $($ $($ '$ '$ tasks, a benefit not available in conventional $ $ software. In other words, it allows the use of non-OS # $ % " "$) ) managed interrupt handlers in a multitasking $($ '$ environment.
    [Show full text]
  • RTA-OSEK Binding Manual: TMS470/TI
    RTA-OSEK Binding Manual: TMS470/TI Contact Details ETAS Group www.etasgroup.com Germany USA ETAS GmbH ETAS Inc. Borsigstraße 14 3021 Miller Road 70469 Stuttgart Ann Arbor, MI 48103 Tel.:+49 (711) 8 96 61-102 Tel.: +1 (888) ETAS INC Fax:+49 (711) 8 96 61-106 Fax: +1 (734) 997-94 49 www.etas.de www.etasinc.com Japan France ETAS K.K. ETAS S.A.S. Queen's Tower C-17F, 1, place des États-Unis 2-3-5, Minatomirai, Nishi-ku, SILIC 307 Yokohama, Kanagawa 94588 Rungis Cedex 220-6217 Japan Tel.: +33 (1) 56 70 00 50 Tel.: +81 (45) 222-0900 Fax: +33 (1) 56 70 00 51 Fax: +81 (45) 222-0956 www.etas.fr www.etas.co.jp Korea Great Britain ETAS Korea Co. Ltd. ETAS UK Ltd. 4F, 705 Bldg. 70-5 Studio 3, Waterside Court Yangjae-dong, Seocho-gu Third Avenue, Centrum 100 Seoul 137-889, Korea Burton-upon-Trent Tel.: +82 (2) 57 47-016 Staffordshire DE14 2WQ Fax: +82 (2) 57 47-120 Tel.: +44 (0) 1283 - 54 65 12 www.etas.co.kr Fax: +44 (0) 1283 - 54 87 67 www.etas-uk.net Copyright Notice © 2001 - 2007 LiveDevices Ltd. All rights reserved. Version: M00088-001 No part of this document may be reproduced without the prior written consent of LiveDevices Ltd. The software described in this document is furnished under a license and may only be used or copied in accordance with the terms of such a license. Disclaimer The information in this document is subject to change without notice and does not represent a commitment on any part of LiveDevices.
    [Show full text]
  • Bringing Virtualization to the X86 Architecture with the Original Vmware Workstation
    12 Bringing Virtualization to the x86 Architecture with the Original VMware Workstation EDOUARD BUGNION, Stanford University SCOTT DEVINE, VMware Inc. MENDEL ROSENBLUM, Stanford University JEREMY SUGERMAN, Talaria Technologies, Inc. EDWARD Y. WANG, Cumulus Networks, Inc. This article describes the historical context, technical challenges, and main implementation techniques used by VMware Workstation to bring virtualization to the x86 architecture in 1999. Although virtual machine monitors (VMMs) had been around for decades, they were traditionally designed as part of monolithic, single-vendor architectures with explicit support for virtualization. In contrast, the x86 architecture lacked virtualization support, and the industry around it had disaggregated into an ecosystem, with different ven- dors controlling the computers, CPUs, peripherals, operating systems, and applications, none of them asking for virtualization. We chose to build our solution independently of these vendors. As a result, VMware Workstation had to deal with new challenges associated with (i) the lack of virtual- ization support in the x86 architecture, (ii) the daunting complexity of the architecture itself, (iii) the need to support a broad combination of peripherals, and (iv) the need to offer a simple user experience within existing environments. These new challenges led us to a novel combination of well-known virtualization techniques, techniques from other domains, and new techniques. VMware Workstation combined a hosted architecture with a VMM. The hosted architecture enabled a simple user experience and offered broad hardware compatibility. Rather than exposing I/O diversity to the virtual machines, VMware Workstation also relied on software emulation of I/O devices. The VMM combined a trap-and-emulate direct execution engine with a system-level dynamic binary translator to ef- ficiently virtualize the x86 architecture and support most commodity operating systems.
    [Show full text]
  • Interrupt Handling
    ,ch10.10847 Page 258 Friday, January 21, 2005 10:54 AM CHAPTER 10 Chapter 10 Interrupt Handling Although some devices can be controlled using nothing but their I/O regions, most real devices are a bit more complicated than that. Devices have to deal with the external world, which often includes things such as spinning disks, moving tape, wires to distant places, and so on. Much has to be done in a time frame that is differ- ent from, and far slower than, that of the processor. Since it is almost always undesir- able to have the processor wait on external events, there must be a way for a device to let the processor know when something has happened. That way, of course, is interrupts. An interrupt is simply a signal that the hardware can send when it wants the processor’s attention. Linux handles interrupts in much the same way that it handles signals in user space. For the most part, a driver need only register a handler for its device’s interrupts, and handle them properly when they arrive. Of course, underneath that simple picture there is some complexity; in particular, interrupt handlers are somewhat limited in the actions they can perform as a result of how they are run. It is difficult to demonstrate the use of interrupts without a real hardware device to generate them. Thus, the sample code used in this chapter works with the parallel port. Such ports are starting to become scarce on modern hardware, but, with luck, most people are still able to get their hands on a system with an available port.
    [Show full text]
  • Interrupt Handling with KSDK and Kinetis Design Studio
    Freescale Semiconductor Interrupt handling with KSDK and Kinetis Design Studio By: Jorge Gonzalez / Technical Information Center Freescale Semiconductor About this document The Kinetis Software Development Kit (KSDK) is intended for rapid evaluation and development with Kinetis family MCUs. Besides of the peripheral drivers, the hardware abstraction layer and middleware stacks, the platform provides a robust interrupt handling mechanism. This document explains the implementation and handling of interrupts when using KSDK in baremetal mode and when using MQX RTOS. Kinetis Design Studio IDE was used as reference, but the concepts should apply for any particular IDE supported by KSDK. Software versions The contents of this document are valid for the latest versions of Kinetis SDK and Kinetis Design Studio by the time of writing, listed below: . KSDK v1.2.0 . KDS v3.0.0 Content 1. GLOSSARY 2. CONCEPTS AND OVERVIEW 2.1 Interrupt Manager 2.2 Vector table location 2.3 Interrupt priorities 3. KSDK INTERRUPT HANDLING 3.1 Baremetal interrupt handling 3.2 MQX RTOS interrupt handling 3.3 Operating System Abstraction layer (OSA) 4. KDS AND PROCESSOR EXPERT CONSIDERATIONS 4.1 KSDK baremetal 4.2 KSDK baremetal + Processor Expert 4.3 MQX for KSDK 4.4 MQX for KSDK + Processor Expert 5. REFERENCES Interrupt handling with KSDK and Kinetis Design Studio 2 Freescale Semiconductor Freescale Semiconductor 1. GLOSSARY KSDK Kinetis Software Development Kit: Set of peripheral drivers, stacks and middleware layers for Kinetis microcontrollers. KDS Kinetis Design Studio: Integrated Development Environment (IDE) software for Kinetis MCUs. API Application Programming Interface: Refers to the set of functions, methods and macros provided by the different layers of the KSDK platform.
    [Show full text]
  • An Architectural Overview of QNX®
    An Architectural Overview of QNX® Dan Hildebrand Quantum Software Systems Ltd. 175 Terrence Matthews Kanata, Ontario K2M 1W8 Canada (613) 591-0931 [email protected] Abstract* This paper presents an architectural overview of the QNX operating system. QNX is an OS that provides applications with a fully network- and multi- processor-distributed, realtime environment that delivers nearly the full, device-level performance of the underlying hardware. The OS architecture used to deliver this operating environment is that of a realtime microkernel surrounded by a collection of optional processes that provide POSIX- and UNIX-compatible system services. By including or excluding various resource managers at run time, QNX can be scaled down for ROM-based embedded systems, or scaled up to encompass hundreds of processorsÅ either tightly or loosely connected by various LAN technologies. Confor- mance to POSIX standard 1003.1, draft standard 1003.2 (shell and utilities) and draft standard 1003.4 (realtime) is maintained transparently throughout the distributed environment. *This paper appeared in the Proceedings of the Usenix Workshop on Micro-Kernels & Other Kernel Architectures, Seattle, April, 1992. ISBN 1-880446-42-1 QNX is a registered trademark and FLEET is a trademark of Quantum Software Systems Ltd. Quantum Software Systems Ltd. An Architectural Overview of QNX® Architecture: Past and Present From its creation in 1982, the QNX architecture has been fundamentally similar to its current formÅthat of a very small microkernel (approximately 10K at that time) surrounded by a team of cooperating processes that provide higher-level OS services. To date, QNX has been used in nearly 200,000 systems, predominantly in applications where realtime performance, development flexibility, and network flexibility have been fundamental requirements.
    [Show full text]
  • AMD APP SDK V2.9.1
    AMD APP SDK v2.9.1 FAQ 1 General Questions 1. Do I need to use additional software with the SDK? To run an OpenCL™ application, you must have an OpenCL™ runtime on your system. If your system includes a recent AMD discrete GPU, or an APU, you also should install the latest Catalyst™ drivers, which can be downloaded from AMD.com. Information on supported devices can be found at developer.amd.com/appsdk. If your system does not include a recent AMD discrete GPU, or APU, the SDK installs a CPU-only OpenCL™ run-time. Also, we recommend using the debugging profiling and analysis tools contained in the AMD CodeXL heterogeneous compute tools suite. 2. Which versions of the OpenCL™ standard does this SDK support? AMD APP SDK 2.9.1 supports the development of applications using the OpenCL™ Specification v 1.2. 3. Will applications developed to execute on OpenCL™ 1.1 still operate in an OpenCL™ 1.2 environment? OpenCL™ is designed to be backwards compatible. The OpenCL™ 1.2 run-time delivered with the AMD Catalyst drivers run any OpenCL™ 1.1-compliant application. However, an OpenCL™ 1.2-compliant application will not execute on an OpenCL™ 1.1 run-time if APIs only supported by OpenCL™ 1.2 are used. 4. Does AMD provide any additional OpenCL™ samples, other than those contained within the SDK? The most recent versions of all of the samples contained within the SDK are also available for individual download from the developer.amd.com/appsdk “Samples & Demos” page. This page also contains additional samples that either were too large to include in the SDK, or which have been developed since the most recent SDK release.
    [Show full text]
  • XOS: an Application-Defined Operating System for Datacenter Computing
    XOS: An Application-Defined Operating System for Datacenter Computing Chen Zheng∗y, Lei Wang∗, Sally A. McKeez, Lixin Zhang∗, Hainan Yex and Jianfeng Zhan∗y ∗State Key Laboratory of Computer Architecture, Institute of Computing Technology, Chinese Academy of Sciences yUniversity of Chinese Academy of Sciences, China zChalmers University of Technology xBeijing Academy of Frontier Sciences and Technology Abstract—Rapid growth of datacenter (DC) scale, urgency of ability, and isolation. Such general-purpose, one-size-fits-all cost control, increasing workload diversity, and huge software designs simply cannot meet the needs of all applications. The investment protection place unprecedented demands on the op- kernel traditionally controls resource abstraction and alloca- erating system (OS) efficiency, scalability, performance isolation, and backward-compatibility. The traditional OSes are not built tion, which hides resource-management details but lengthens to work with deep-hierarchy software stacks, large numbers of application execution paths. Giving applications control over cores, tail latency guarantee, and increasingly rich variety of functionalities usually reserved for the kernel can streamline applications seen in modern DCs, and thus they struggle to meet systems, improving both individual performances and system the demands of such workloads. throughput [5]. This paper presents XOS, an application-defined OS for modern DC servers. Our design moves resource management There is a large and growing gap between what DC ap- out of the
    [Show full text]
  • Operating Systems
    Operating Systems − they will be added as we go • you can also use slides from previous years Petr Ročkai − they are already in study materials − but: not everything is covered in those These are draft (work in progress) lecture notes for PB152. It is expected to be gradually ��illed in, hopefully before the semester You are reading the lecture notes for the course PB152 Operating ends. For now, it is mostly a printable version of the slides. Systems. This is a supplementary resource based on the lecture Organisation slides, but with additional details that would not it into the slide format. These lecture notes should be self-contained in the sense • lectures, with an optional seminar that they only rely on knowledge you have from other courses, • written exam at the end like PB150 Computer Systems (or PB151) or PB071 Principles of − multiple choice Low-Level programming. Likewise, being familiar with the topics − free-form questions covered in these lecture notes is suficient to pass the exam. • 1 online test mid-term, 1 before exam − mainly training for the exam proper Books The written exam will consist of two parts, a multiple-choice test • there are a few good OS books (with exactly one correct answer on each question) and a free- • you are encouraged to get and read them form question. Additionally, you will be required to pass a mid- term test with the same pool of multiple-choice questions. The • A. Tanenbaum: Modern Operating Systems question pool is known in advance: both the entire mid-term and • A.
    [Show full text]