Machine Learning for Load Balancing in the Linux Kernel Jingde Chen Subho S

Total Page:16

File Type:pdf, Size:1020Kb

Machine Learning for Load Balancing in the Linux Kernel Jingde Chen Subho S Machine Learning for Load Balancing in the Linux Kernel Jingde Chen Subho S. Banerjee [email protected] [email protected] University of Illinois at Urbana-Champaign University of Illinois at Urbana-Champaign Zbigniew T. Kalbarczyk Ravishankar K. Iyer [email protected] [email protected] University of Illinois at Urbana-Champaign University of Illinois at Urbana-Champaign Abstract ACM Reference Format: The OS load balancing algorithm governs the performance Jingde Chen, Subho S. Banerjee, Zbigniew T. Kalbarczyk, and Rav- gains provided by a multiprocessor computer system. The ishankar K. Iyer. 2020. Machine Learning for Load Balancing in Linux’s Completely Fair Scheduler (CFS) scheduler tracks the Linux Kernel. In 11th ACM SIGOPS Asia-Pacific Workshop on Systems (APSys ’20), August 24–25, 2020, Tsukuba, Japan. ACM, New process loads by average CPU utilization to balance work- York, NY, USA, 8 pages. https://doi.org/10.1145/3409963.3410492 load between processor cores. That approach maximizes the utilization of processing time but overlooks the con- tention for lower-level hardware resources. In servers run- 1 Introduction ning compute-intensive workloads, an imbalanced need for The Linux scheduler, called the Completely Fair Scheduler limited computing resources hinders execution performance. (CFS) algorithm, enforces a fair scheduling policy that aims to This paper solves the above problem using a machine learn- allow all tasks to be executed with a fair quota of processing ing (ML)-based resource-aware load balancer. We describe time. CFS allocates CPU time appropriately among the avail- (1) low-overhead methods for collecting training data; (2) an able hardware threads to allow processes to make forward ML model based on a multi-layer perceptron model that imi- progress. However, the scheduling algorithm of CFS does tates the CFS load balancer based on the collected training not take into account contention for low-level hardware re- data; and (3) an in-kernel implementation of inference on sources (e.g., shared functional units or memory bandwidth) the model. Our experiments demonstrate that the proposed other than CPU time. There is an implicit assumption that model has an accuracy of 99% in making migration decisions fair sharing of CPU time implies that other such microarchi- and while only increasing the latency by 1.9 µs. tectural resources are also divided fairly [2]. That assumption CCS Concepts is often false and leaves the scheduling algorithm unaware • Software and its engineering ! Scheduling; • Theory of hardware resource utilization, thereby leading to shared of computation ! Scheduling algorithms; • Comput- resource contention and reduced performance. ing methodologies ! Neural networks. While prior work has investigated the integration of hard- ware resource usage in OS load balancing, most attempts Keywords have focused only on resources shared at the chip multi- Linux kernel, Machine Learning, Neural Network, Com- processor (CMP) level and depend on predefined schedul- pletely Fair Scheduler, Operating System, Load Balancing. ing policies or on efforts by the user to manually annotate resource usage of the programs [5, 7, 16]. There is not a Permission to make digital or hard copies of all or part of this work for system-wide solution that works for all general processes, personal or classroom use is granted without fee provided that copies or one that can take advantage of the dynamic contextual are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights information available from tracing and telemetry data from for components of this work owned by others than the author(s) must the system. Traditional methods are fundamentally limited be honored. Abstracting with credit is permitted. To copy otherwise, or in having to effectively quantify relationships between differ- republish, to post on servers or to redistribute to lists, requires prior specific ent resources and application performance, which indeed is permission and/or a fee. Request permissions from [email protected]. a hard problem. One solution is to use machine learning (ML) APSys ’20, August 24–25, 2020, Tsukuba, Japan in the kernel to model and tackle the problems described © 2020 Copyright held by the owner/author(s). Publication rights licensed to ACM. above. Even small gains can translate into big savings at the ACM ISBN 978-1-4503-8069-0/20/08...$15.00 level of large data centers. An important step in building an https://doi.org/10.1145/3409963.3410492 ML engine that can outperform the standard CFS Linux load APSys ’20, August 24–25, 2020, Tsukuba, Japan Jingde Chen, Subho S. Banerjee, Zbigniew T. Kalbarczyk, and Ravishankar K. Iyer balancer is to see whether it is possible to match the current kernel had used a straightforward $ ¹=º scheduling algo- load balancer performance. The insights thus gained, as we rithm, which iterated every runnable task to pick the next will show, can be quite valuable in the next step, namely one to run. In 2003, the $ ¹=º scheduler was replaced with building an engine that can outperform current techniques. the $ ¹1º scheduler which runs in constant time instead of ML in the Kernel. The ML approach will have to (i) per- linear time. The $ ¹1º scheduler managed running tasks with form load balancing in Linux that emulates original kernel two linked lists for each task priority, one for ready-to-run decisions; (ii) include hardware resource utilization models; tasks and the other for expired tasks that have used up their (iii) train the model to solve specific cases with high resource timeshares. Later, Linux abandoned the concept of having needs that cause frequent contention; and (iv) apply rein- predefined time slices. The Completely Fair Scheduler (CFS) forcement learning to the model to optimize load balancing was introduced to the kernel in Linux 2.6.23 in 2007 together performance. To exemplify the types of resource contention, with the design of the Modular Scheduler Core [13], and has consider simultaneous multi-threaded (SMT) threads on a since been the default Linux scheduler. We refer to that latest CMP that appear as identical logical CPU cores to the OS, design as the Linux scheduler. The scheduler implements in- but share most of the computing resources on the chip, in- dividual scheduling classes to which the Modular Scheduler cluding caches and registers. Processes scheduled to run on Core can delegate detailed scheduling tasks. Existing sched- different SMT threads at the same time contend for there- uling classes include the rt class for real-time processes; the sources on the chip. Similarly, CPUs within the same NUMA fair class which implements the CFS; and the idle class, node share memory and I/O bandwidth. In the end, we aim which handles cases when there is no runnable task. to implement an ML module inside the kernel that monitors The scheduler comprises a large portion of the kernel real-time resource usage in the system, identifies potential source and resides in the <kernel/sched> directory and the hardware performance bottlenecks, and generates the best scheduler core is implemented in <kernel/sched/core.c>. load balancing decisions. The Linux scheduler uses per-CPU allocated runqueues for Contributions. In this paper, we present the results of the faster local access, and the load balancing algorithm works by first step, i.e., an ML model and an implementation to replace migrating tasks across the per-core runqueues. The primary the CFS load balancer. The proposed approach consists of: granularity of work scheduled under Linux is the thread (and (1) Collection of runtime system data via dynamic kernel not the process). The threads are not scheduled via direct tracing, using eBPF (described in Section 3.2). attachment to the per-core runqueues, but as schedulable (2) Training of the ML model using collected data to emulate entities (sched_entity) attached to subordinate class run- task migration decisions (described in Section 3.3). queues managed by individual scheduling classes. (3) Two implementations of the model in the kernel, one 2.1 The Completely Fair Scheduler with floating-point arithmetic and one with fixed-point The CFS is the most complicated part of the Linux sched- arithmetic (described in Section 3.4). uler and is implemented in <kernel/sched/fair.c> with The ML model is trained offline in the user space andis 10,000 lines of code. The CFS schedules threads that have used for online inference in the kernel to generate migration policies of SCHED_NORMAL and SCHED_BATCH. As mentioned decisions. The performance of the in-kernel ML model is as above, the CFS manages its own runqueue structure cfs_rq follows: in the per-CPU runqueue. cfs_rq sorts the sched_entity (1) 99% of the migration decisions generated by the ML of all runnable threads on it in a red-black tree (a kind of model are correct. self-balancing binary search tree) that uses virtual runtimes (2) The average latency of the load balancing procedure is as keys. When asked to pick the next task to run, CFS returns 13% higher than the original kernel (16.4 µs vs. 14.5 µs). the task corresponding to the sched_entity with the most (3) During testing, our approach produced a 5-minute aver- insignificant virtual runtime, which is the least executed task. age CPU load of 1.773 compared to Linux’s 1.758. In that way, all tasks in the queue can get a fair chance to The work described in this paper was based on the Linux run. To reflect task priorities, the virtual runtime accumu- 4.15 source code. The code used in this paper can be found at lates more slowly for high-priority tasks and they get to https://github.com/keitokuch/MLLB, and the modified ker- run longer.
Recommended publications
  • Proceedings 2005
    LAC2005 Proceedings 3rd International Linux Audio Conference April 21 – 24, 2005 ZKM | Zentrum fur¨ Kunst und Medientechnologie Karlsruhe, Germany Published by ZKM | Zentrum fur¨ Kunst und Medientechnologie Karlsruhe, Germany April, 2005 All copyright remains with the authors www.zkm.de/lac/2005 Content Preface ............................................ ............................5 Staff ............................................... ............................6 Thursday, April 21, 2005 – Lecture Hall 11:45 AM Peter Brinkmann MidiKinesis – MIDI controllers for (almost) any purpose . ....................9 01:30 PM Victor Lazzarini Extensions to the Csound Language: from User-Defined to Plugin Opcodes and Beyond ............................. .....................13 02:15 PM Albert Gr¨af Q: A Functional Programming Language for Multimedia Applications .........21 03:00 PM St´ephane Letz, Dominique Fober and Yann Orlarey jackdmp: Jack server for multi-processor machines . ......................29 03:45 PM John ffitch On The Design of Csound5 ............................... .....................37 04:30 PM Pau Arum´ıand Xavier Amatriain CLAM, an Object Oriented Framework for Audio and Music . .............43 Friday, April 22, 2005 – Lecture Hall 11:00 AM Ivica Ico Bukvic “Made in Linux” – The Next Step .......................... ..................51 11:45 AM Christoph Eckert Linux Audio Usability Issues .......................... ........................57 01:30 PM Marije Baalman Updates of the WONDER software interface for using Wave Field Synthesis . 69 02:15 PM Georg B¨onn Development of a Composer’s Sketchbook ................. ....................73 Saturday, April 23, 2005 – Lecture Hall 11:00 AM J¨urgen Reuter SoundPaint – Painting Music ........................... ......................79 11:45 AM Michael Sch¨uepp, Rene Widtmann, Rolf “Day” Koch and Klaus Buchheim System design for audio record and playback with a computer using FireWire . 87 01:30 PM John ffitch and Tom Natt Recording all Output from a Student Radio Station .
    [Show full text]
  • Oracle® Linux 7 Monitoring and Tuning the System
    Oracle® Linux 7 Monitoring and Tuning the System F32306-03 October 2020 Oracle Legal Notices Copyright © 2020, Oracle and/or its affiliates. This software and related documentation are provided under a license agreement containing restrictions on use and disclosure and are protected by intellectual property laws. Except as expressly permitted in your license agreement or allowed by law, you may not use, copy, reproduce, translate, broadcast, modify, license, transmit, distribute, exhibit, perform, publish, or display any part, in any form, or by any means. Reverse engineering, disassembly, or decompilation of this software, unless required by law for interoperability, is prohibited. The information contained herein is subject to change without notice and is not warranted to be error-free. If you find any errors, please report them to us in writing. If this is software or related documentation that is delivered to the U.S. Government or anyone licensing it on behalf of the U.S. Government, then the following notice is applicable: U.S. GOVERNMENT END USERS: Oracle programs (including any operating system, integrated software, any programs embedded, installed or activated on delivered hardware, and modifications of such programs) and Oracle computer documentation or other Oracle data delivered to or accessed by U.S. Government end users are "commercial computer software" or "commercial computer software documentation" pursuant to the applicable Federal Acquisition Regulation and agency-specific supplemental regulations. As such, the use, reproduction, duplication, release, display, disclosure, modification, preparation of derivative works, and/or adaptation of i) Oracle programs (including any operating system, integrated software, any programs embedded, installed or activated on delivered hardware, and modifications of such programs), ii) Oracle computer documentation and/or iii) other Oracle data, is subject to the rights and limitations specified in the license contained in the applicable contract.
    [Show full text]
  • Chapter 3. Booting Operating Systems
    Chapter 3. Booting Operating Systems Abstract: Chapter 3 provides a complete coverage on operating systems booting. It explains the booting principle and the booting sequence of various kinds of bootable devices. These include booting from floppy disk, hard disk, CDROM and USB drives. Instead of writing a customized booter to boot up only MTX, it shows how to develop booter programs to boot up real operating systems, such as Linux, from a variety of bootable devices. In particular, it shows how to boot up generic Linux bzImage kernels with initial ramdisk support. It is shown that the hard disk and CDROM booters developed in this book are comparable to GRUB and isolinux in performance. In addition, it demonstrates the booter programs by sample systems. 3.1. Booting Booting, which is short for bootstrap, refers to the process of loading an operating system image into computer memory and starting up the operating system. As such, it is the first step to run an operating system. Despite its importance and widespread interests among computer users, the subject of booting is rarely discussed in operating system books. Information on booting are usually scattered and, in most cases, incomplete. A systematic treatment of the booting process has been lacking. The purpose of this chapter is to try to fill this void. In this chapter, we shall discuss the booting principle and show how to write booter programs to boot up real operating systems. As one might expect, the booting process is highly machine dependent. To be more specific, we shall only consider the booting process of Intel x86 based PCs.
    [Show full text]
  • Inside the Linux 2.6 Completely Fair Scheduler
    Inside the Linux 2.6 Completely Fair Scheduler http://www.ibm.com/developerworks/library/l-completely-fair-scheduler/ developerWorks Technical topics Linux Technical library Inside the Linux 2.6 Completely Fair Scheduler Providing fair access to CPUs since 2.6.23 The task scheduler is a key part of any operating system, and Linux® continues to evolve and innovate in this area. In kernel 2.6.23, the Completely Fair Scheduler (CFS) was introduced. This scheduler, instead of relying on run queues, uses a red-black tree implementation for task management. Explore the ideas behind CFS, its implementation, and advantages over the prior O(1) scheduler. M. Tim Jones is an embedded firmware architect and the author of Artificial Intelligence: A Systems Approach, GNU/Linux Application Programming (now in its second edition), AI Application Programming (in its second edition), and BSD Sockets Programming from a Multilanguage Perspective. His engineering background ranges from the development of kernels for geosynchronous spacecraft to embedded systems architecture and networking protocols development. Tim is a Consultant Engineer for Emulex Corp. in Longmont, Colorado. 15 December 2009 Also available in Japanese The Linux scheduler is an interesting study in competing pressures. On one side are the use models in which Linux is applied. Although Linux was originally developed as a desktop operating system experiment, you'll now find it on servers, tiny embedded devices, mainframes, and supercomputers. Not surprisingly, the scheduling loads for these domains differ. On the other side are the technological advances made in the platform, including architectures (multiprocessing, symmetric multithreading, non-uniform memory access [NUMA]) and virtualization.
    [Show full text]
  • UNIX OS Agent User's Guide
    IBM Tivoli Monitoring Version 6.3.0 UNIX OS Agent User's Guide SC22-5452-00 IBM Tivoli Monitoring Version 6.3.0 UNIX OS Agent User's Guide SC22-5452-00 Note Before using this information and the product it supports, read the information in “Notices” on page 399. This edition applies to version 6, release 3 of IBM Tivoli Monitoring (product number 5724-C04) and to all subsequent releases and modifications until otherwise indicated in new editions. © Copyright IBM Corporation 1994, 2013. US Government Users Restricted Rights – Use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM Corp. Contents Tables ...............vii Solaris System CPU Workload workspace ....28 Solaris Zone Processes workspace .......28 Chapter 1. Using the monitoring agent . 1 Solaris Zones workspace ..........28 System Details workspace .........28 New in this release ............2 System Information workspace ........29 Components of the monitoring agent ......3 Top CPU-Memory %-VSize Details workspace . 30 User interface options ...........4 UNIX OS workspace ...........30 UNIX Detail workspace ..........31 Chapter 2. Requirements for the Users workspace ............31 monitoring agent ...........5 Enabling the Monitoring Agent for UNIX OS to run Chapter 4. Attributes .........33 as a nonroot user .............7 Agent Availability Management Status attributes . 36 Securing your IBM Tivoli Monitoring installation 7 Agent Active Runtime Status attributes .....37 Setting overall file ownership and permissions for AIX AMS attributes............38
    [Show full text]
  • SUSE Linux Enterprise Server 11 SP4 System Analysis and Tuning Guide System Analysis and Tuning Guide SUSE Linux Enterprise Server 11 SP4
    SUSE Linux Enterprise Server 11 SP4 System Analysis and Tuning Guide System Analysis and Tuning Guide SUSE Linux Enterprise Server 11 SP4 Publication Date: September 24, 2021 SUSE LLC 1800 South Novell Place Provo, UT 84606 USA https://documentation.suse.com Copyright © 2006– 2021 SUSE LLC and contributors. All rights reserved. Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or (at your option) version 1.3; with the Invariant Section being this copyright notice and license. A copy of the license version 1.2 is included in the section entitled “GNU Free Documentation License”. For SUSE trademarks, see http://www.suse.com/company/legal/ . All other third party trademarks are the property of their respective owners. A trademark symbol (®, ™ etc.) denotes a SUSE or Novell trademark; an asterisk (*) denotes a third party trademark. All information found in this book has been compiled with utmost attention to detail. However, this does not guarantee complete accuracy. Neither SUSE LLC, its aliates, the authors nor the translators shall be held liable for possible errors or the consequences thereof. Contents About This Guide xi 1 Available Documentation xii 2 Feedback xiv 3 Documentation Conventions xv I BASICS 1 1 General Notes on System Tuning 2 1.1 Be Sure What Problem to Solve 2 1.2 Rule Out Common Problems 3 1.3 Finding the Bottleneck 3 1.4 Step-by-step Tuning 4 II SYSTEM MONITORING 5 2 System Monitoring Utilities 6 2.1 Multi-Purpose Tools 6 vmstat 7
    [Show full text]
  • System Analysis and Tuning Guide System Analysis and Tuning Guide SUSE Linux Enterprise Server 15 SP1
    SUSE Linux Enterprise Server 15 SP1 System Analysis and Tuning Guide System Analysis and Tuning Guide SUSE Linux Enterprise Server 15 SP1 An administrator's guide for problem detection, resolution and optimization. Find how to inspect and optimize your system by means of monitoring tools and how to eciently manage resources. Also contains an overview of common problems and solutions and of additional help and documentation resources. Publication Date: September 24, 2021 SUSE LLC 1800 South Novell Place Provo, UT 84606 USA https://documentation.suse.com Copyright © 2006– 2021 SUSE LLC and contributors. All rights reserved. Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or (at your option) version 1.3; with the Invariant Section being this copyright notice and license. A copy of the license version 1.2 is included in the section entitled “GNU Free Documentation License”. For SUSE trademarks, see https://www.suse.com/company/legal/ . All other third-party trademarks are the property of their respective owners. Trademark symbols (®, ™ etc.) denote trademarks of SUSE and its aliates. Asterisks (*) denote third-party trademarks. All information found in this book has been compiled with utmost attention to detail. However, this does not guarantee complete accuracy. Neither SUSE LLC, its aliates, the authors nor the translators shall be held liable for possible errors or the consequences thereof. Contents About This Guide xii 1 Available Documentation xiii
    [Show full text]
  • Thread Scheduling in Multi-Core Operating Systems Redha Gouicem
    Thread Scheduling in Multi-core Operating Systems Redha Gouicem To cite this version: Redha Gouicem. Thread Scheduling in Multi-core Operating Systems. Computer Science [cs]. Sor- bonne Université, 2020. English. tel-02977242 HAL Id: tel-02977242 https://hal.archives-ouvertes.fr/tel-02977242 Submitted on 24 Oct 2020 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Ph.D thesis in Computer Science Thread Scheduling in Multi-core Operating Systems How to Understand, Improve and Fix your Scheduler Redha GOUICEM Sorbonne Université Laboratoire d’Informatique de Paris 6 Inria Whisper Team PH.D.DEFENSE: 23 October 2020, Paris, France JURYMEMBERS: Mr. Pascal Felber, Full Professor, Université de Neuchâtel Reviewer Mr. Vivien Quéma, Full Professor, Grenoble INP (ENSIMAG) Reviewer Mr. Rachid Guerraoui, Full Professor, École Polytechnique Fédérale de Lausanne Examiner Ms. Karine Heydemann, Associate Professor, Sorbonne Université Examiner Mr. Etienne Rivière, Full Professor, University of Louvain Examiner Mr. Gilles Muller, Senior Research Scientist, Inria Advisor Mr. Julien Sopena, Associate Professor, Sorbonne Université Advisor ABSTRACT In this thesis, we address the problem of schedulers for multi-core architectures from several perspectives: design (simplicity and correct- ness), performance improvement and the development of application- specific schedulers.
    [Show full text]
  • Package Manager: the Core of a GNU/Linux Distribution
    Package Manager: The Core of a GNU/Linux Distribution by Andrey Falko A Thesis submitted to the Faculty in partial fulfillment of the requirements for the BACHELOR OF ARTS Accepted Paul Shields, Thesis Adviser Allen Altman, Second Reader Christopher K. Callanan, Third Reader Mary B. Marcy, Provost and Vice President Simon’s Rock College Great Barrington, Massachusetts 2007 Abstract Package Manager: The Core of a GNU/Linux Distribution by Andrey Falko As GNU/Linux operating systems have been growing in popularity, their size has also been growing. To deal with this growth people created special programs to organize the software available for GNU/Linux users. These programs are called package managers. There exists a very wide variety of package managers, all offering their own benefits and deficiencies. This thesis explores all of the major aspects of package management in the GNU/Linux world. It covers what it is like to work without package managers, primitive package man- agement techniques, and modern package management schemes. The thesis presents the creation of a package manager called Vestigium. The creation of Vestigium is an attempt to automate the handling of file collisions between packages, provide a seamless interface for installing both binary packages and packages built from source, and to allow per package optimization capabilities. Some of the features Vestigium is built to have are lacking in current package managers. No current package manager contains all the features which Vestigium is built to have. Additionally, the thesis explains the problems that developers face in maintaining their respective package managers. i Acknowledgments I thank my thesis committee members, Paul Shields, Chris Callanan, and Allen Altman for being patient with my error-ridden drafts.
    [Show full text]
  • Linux CPU Schedulers: CFS and Muqss Comparison
    Umeå University 2021-06-14 Bachelor’s Thesis In Computing Science Linux CPU Schedulers: CFS and MuQSS Comparison Name David Shakoori Gustafsson Supervisor Anna Jonsson Examinator Ola Ringdahl CFS and MuQSS Comparison Abstract The goal of this thesis is to compare two process schedulers for the Linux operating system. In order to provide a responsive and interactive user ex- perience, an efficient process scheduling algorithm is important. This thesis seeks to explain the potential performance differences by analysing the sched- ulers’ respective designs. The two schedulers that are tested and compared are Con Kolivas’s MuQSS and Linux’s default scheduler, CFS. They are tested with respect to three main aspects: latency, turn-around time and interactivity. Latency is tested by using benchmarking software, the turn- around time by timing software compilation, and interactivity by measuring video frame drop percentages under various background loads. These tests are performed on a desktop PC running Linux OpenSUSE Leap 15.2, using kernel version 5.11.18. The test results show that CFS manages to keep a generally lower latency, while turn-around times differs little between the two. Running the turn-around time test’s compilation using a single process gives MuQSS a small advantage, while dividing the compilation evenly among the available logical cores yields little difference. However, CFS clearly outper- forms MuQSS in the interactivity test, where it manages to keep frame drop percentages considerably lower under each tested background load. As is apparent by the results, Linux’s current default scheduler provides a more responsive and interactive experience within the testing conditions, than the alternative MuQSS.
    [Show full text]
  • Linux-Cookbook.Pdf
    LINUX COOKBOOK ™ Other Linux resources from O’Reilly Related titles Linux Device Drivers Exploring the JDS Linux Linux in a Nutshell Desktop Running Linux Learning Red Hat Enterprise Building Embedded Linux Linux and Fedora Systems Linux Pocket Guide Linux Security Cookbook Understanding the Linux Kernel Linux Books linux.oreilly.com is a complete catalog of O’Reilly’s books on Resource Center Linux and Unix and related technologies, including sample chapters and code examples. ONLamp.com is the premier site for the open source web plat- form: Linux, Apache, MySQL, and either Perl, Python, or PHP. Conferences O’Reilly brings diverse innovators together to nurture the ideas that spark revolutionary industries. We specialize in document- ing the latest tools and systems, translating the innovator’s knowledge into useful skills for those in the trenches. Visit conferences.oreilly.com for our upcoming events. Safari Bookshelf (safari.oreilly.com) is the premier online refer- ence library for programmers and IT professionals. Conduct searches across more than 1,000 books. Subscribers can zero in on answers to time-critical questions in a matter of seconds. Read the books on your Bookshelf from cover to cover or sim- ply flip to the page you need. Try it today with a free trial. LINUX COOKBOOK ™ Carla Schroder Beijing • Cambridge • Farnham • Köln • Paris • Sebastopol • Taipei • Tokyo Linux Cookbook™ by Carla Schroder Copyright © 2005 O’Reilly Media, Inc. All rights reserved. Printed in the United States of America. Published by O’Reilly Media, Inc., 1005 Gravenstein Highway North, Sebastopol, CA 95472. O’Reilly books may be purchased for educational, business, or sales promotional use.
    [Show full text]
  • Evolutions in Recent Hardware and Applications to Virtualization Clementine Maurice
    Information leakage on shared hardware : evolutions in recent hardware and applications to virtualization Clementine Maurice To cite this version: Clementine Maurice. Information leakage on shared hardware : evolutions in recent hardware and applications to virtualization. Cryptography and Security [cs.CR]. Télécom ParisTech, 2015. English. NNT : 2015ENST0070. tel-01439872 HAL Id: tel-01439872 https://pastel.archives-ouvertes.fr/tel-01439872 Submitted on 18 Jan 2017 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. 2015-ENST-0070 EDITE - ED 130 Doctorat ParisTech THÈSE pour obtenir le grade de docteur délivré par TELECOM ParisTech Spécialité « Informatique » présentée et soutenue publiquement par Clémentine MAURICE le 28 octobre 2015 Fuites d’information dans les processeurs récents et applications à la virtualisation Directeurs de thèse : Aurélien FRANCILLON Christoph NEUMANN Co-encadrement de la thèse : Olivier HEEN Jury M. Thomas JENSEN, Directeur de recherche, Inria Rennes, France Président de jury M. Gildas AVOINE, Professeur, IRISA et INSA Rennes,
    [Show full text]