Automating Virtual Machine Network Profiles

Total Page:16

File Type:pdf, Size:1020Kb

Automating Virtual Machine Network Profiles Automating Virtual Machine Network Profiles Vivek Kashyap Arnd Bergman Stefan Berger IBM IBM IBM [email protected] [email protected] [email protected] Gerhard Stenzel Jens Osterkamp IBM IBM [email protected] [email protected] Abstract and VM to external network communication a virtual bridge is included in the Linux host. The Linux vir- With the explosion of use of virtual machines in the tual bridge relays unicast traffic between the virtual and data-center/cloud environments there is correspondingly physical interfaces attached to it. It further replicates a requirement for automating the associated network and forwards broadcast and multicast transmission re- management and administration. The virtual machines ceived on its port(s). share the limited number of network adapters on the sys- For network isolation and security, the iptables/ebtables tem among them but may run workloads with contend- rules might be enforced on the system. The more com- ing network requirements. Furthermore, these work- mon rules are to prevent ARP or IP spoofing, block loads may be run on behalf of customers desiring com- sending a link level broadcast, or to allow only specific plete isolation of their network traffic. The enforcement set of protocol traffic. of network traffic isolation through access controls (fil- ters) and VLANs on the host adds additional run-time Different workloads running in the VMs might also be and administrative overhead. This is further exacerbated specifically restricted within certain limits based on the when Virtual Machines are migrated to another physical workload requirements, priorities and possibly based on system as the corresponding network profiles must be the bandwidth purchased by the user. re-enforced on the target system. The physical switches must also be reprogrammed. For the purposes of this paper the layer-2 forwarding, multiplexing and filtering(ebtables) function is collec- This paper describes the Linux enhancements in kernel, tively considered part of the virtual-bridge. in libvirt and layer-2 networking, enabling the offload- ing of the switching function to the external physical The per-packet processing for bridging function - packet switches while retaining the control in Linux host. The relaying, replicating, evaluation against filter rules, or layer 2 network filters and QoS profiles are automati- bandwidth control - imposes a heavy burden that takes cally migrated with the virtual machine on to the target away CPU resources that could be utilized more gain- system and imposed on the physical switch port without fully. This problem gets more and more exacerbated as administrative intervention. the number of virtual machines supported on a single host increases in the multi-core, and large memory sys- We discuss the proposed IEEE standard (802.1Qbg) and tems being used in the data-center/cloud environments. its implementation on Linux for automated migration of port profiles when a VM is migrated from one system to Another problem faced in large deployments is the need another. for maintaining consistent view of the port profiles. The switch fabric and the embedded switches in the hyper- visor(Linux KVM host) may have different capabilities 1 Introduction and also be under separate administrative control. A network adapter is shared across multiple virtual ma- This is further exacerbated as the VMs (running im- chines(VM) on a Linux host. To accommodate VM-VM portant workloads) migrate from one system to another. • 147 • 148 • Automating Virtual Machine Network Profiles The filter rules and QoS enforcement must follow the Virtual Ethernet Port Aggregator VM and be quickly in place. Not all policies may be deployable in the hypervisor and the switch must be in- The component in the physical end-station that works formed of the migration. together with the adjacent switch to support EVB is called "Virtual Ethernet Port Aggregator" or VEPA. As the MAC address associated with the virtual inter- faces migrates as well, the physical switch at the tar- get does not know if the MAC, now visible at another VM Edge Switch Edge port, is a migrated VM or a different VM using the Regular vSwitch VM same MAC. Thereby, it needs to be informed of the port- mode allows VM- to-VM traffic with VM profile policy to deploy. limited policy VEB L2 net(s) enforcement. VM The proposal ’Edge Virtual Bridging’ in IEEE VM [802.1Qbg] addresses these issues through offloading of Port Profile the switching function to the adjacent bridge i.e. the Database physical switch in the network. vSwitch in VEPA mode forces all VM traffic to fully- capable edge for VM V E L2 net(s) full policy P 1.1 Edge Virtual Bridging: IEEE 802.1Qbg A enforcement. VM VM The Edge Virtual Bridging(EVB) proposal defines the protocols, configuration and control required across the physical end station, such as the Linux host, and the ad- jacent switch. A VEPA is very simplified version of an Ethernet bridge This section provides an introduction to the IEEE that allows multiple downlink ports to communicate 802.1Qbg proposal and the subsequent sections describe with a single uplink port but not with each other. Eth- our implementation of the same on Linux to configure ernet frames from one of the downlink ports get sent di- KVM guests in ’VEPA’ mode. rectly to the uplink, and Ethernet frames arriving at the uplink port get forwarded to just the destination with Reflective Relay the matching MAC address, or flooded to all downlink ports in case of broadcast. A VEPA does not support The proposed standard extends the adjacent bridge ca- unknown unicast frames, which get silently dropped. pabilities to the virtual machines by ensuring that all the packets sent by the VM’s are first sent to the switch port. The VEPA component therefore, is able to provide VM The switch then consults its tables, and if the packet to VM communication in conjunction with an uplink is destined to another VM on the same host, sends the port which is configured in ’reflective relay’ mode. packet back to the host. The packet is then forwarded to the destination VM. Dynamic discovery and configuration for VEPA This packet flow enables the switch to enforce fabric mode wide packet filtering and control policies across all net- work traffic. Otherwise the policy and rules might need The Link-level Discovery Protocol(LLDP)[LLDP] is to be co-ordinated and enforced across the switch and used for layer-2 discovery operations. The EVB pro- the hypervisor leading to administrative overhead and posal extends the TLVs (configuration messages) sup- inefficiencies outlined earlier. ported to include advertisement of ’reflective relay’ ca- pabilities and setting of the adjacent bridge’s port in re- Existent switches do not reflect the packet back on the flective relay mode. The TLV also includes additional port on which it was received. Therefore, a new mode, parameters such as the the number of virtual interfaces referred to as the ’reflective relay’ mode is introduced in (VSI or Virtual station interfaces in EVB parlance) that the Edge Virtual Bridging(EVB) proposal. a station or bridge can support. See [802.1Qbg] for de- scription of all capabilities and parameters exchanged. 2010 Linux Symposium • 149 accordingly. This mechanism also addresses the issue of ’reincarnated’ or reused MAC address and a MAC address that has appeared on the port after a VM migra- Station (e.g., Hypervisor) Bridge tion since the switch is informed of the exact profile to Bridge advertises Server what modes it can configures EVB TLV – OFFER CAPABILTIES 1 support and the max impose. itself from the Capabilities number of VSIs it can available Forwarding: Std, RR handle. capabilities Other: VSI, Auth, etc. according to Current Config.(Std, None) local policy. # VSIs Supported = J # VSIs Configured = 0 The VSI Discovery protocol (VDP) defined by EVB, RTE = 15 therefore defines a set of states and commands ex- 2 EVB TLV - CONFIGURE Bridge Capabilities & Current Config. matches its Forwarding: RR configuration changed between the bridge and the station. These are: Other: VSI, Auth, etc. to the limited # VSIs Supported = J capabilities # VSIs Configured = K advertised by RTE = 10 the station. EVB TLV – CONFIRMATION 3 • Pre-associate: Inform the switch of the VSI and Capabilities Forwarding: Std, RR But still port-profile and intention to associate Other: VSI, Auth, etc. advertises its Current Config. full set of Forwarding: RR capabilities. Other: VSI, Auth, etc. # VSIs Supported = J # VSIs Configured = K • Pre-associate with Resource Reservation: Same as RTE 10 Pre-associate except also reserve resources at the switch for a future association. The bridge periodically advertises its capabilities. The • Associate: Associate the VSI. This causes all re- Station receives the TLV, and based on its configuration, sources to be allocated at the switch and the impo- may respond to configure the port in ’reflective relay’ sition of the VSI port profile. (RR) mode. • DeAssociate: De-associate the VSI from the port VSI Discovery and Configuration and profile The switching function is offloaded from the end-station with VEPA and dynamic configuration of the switch Edge Control Protocol port to RR mode. This however does not fully address The discovery and exchange of bridge capabilities is the problems outlined above. A mechanism is required performed over LLDP. LLDP is an unacknowledged to associate the virtual interface (VSI) to specific profile protocol and has limitations on frequency and number for filtering, VLAN and bandwidth control. of transmissions. This is achieved by informing the switch of the MAC For VDP, the EVB has proposed a new link-level trans- address and VLAN ID pair and the profile that must be port called the "Edge Control Protocol"(ECP) which ac- used.
Recommended publications
  • Flexible Lustre Management
    Flexible Lustre management Making less work for Admins ORNL is managed by UT-Battelle for the US Department of Energy How do we know Lustre condition today • Polling proc / sysfs files – The knocking on the door model – Parse stats, rpc info, etc for performance deviations. • Constant collection of debug logs – Heavy parsing for common problems. • The death of a node – Have to examine kdumps and /or lustre dump Origins of a new approach • Requirements for Linux kernel integration. – No more proc usage – Migration to sysfs and debugfs – Used to configure your file system. – Started in lustre 2.9 and still on going. • Two ways to configure your file system. – On MGS server run lctl conf_param … • Directly accessed proc seq_files. – On MSG server run lctl set_param –P • Originally used an upcall to lctl for configuration • Introduced in Lustre 2.4 but was broken until lustre 2.12 (LU-7004) – Configuring file system works transparently before and after sysfs migration. Changes introduced with sysfs / debugfs migration • sysfs has a one item per file rule. • Complex proc files moved to debugfs • Moving to debugfs introduced permission problems – Only debugging files should be their. – Both debugfs and procfs have scaling issues. • Moving to sysfs introduced the ability to send uevents – Item of most interest from LUG 2018 Linux Lustre client talk. – Both lctl conf_param and lctl set_param –P use this approach • lctl conf_param can set sysfs attributes without uevents. See class_modify_config() – We get life cycle events for free – udev is now involved. What do we get by using udev ? • Under the hood – uevents are collect by systemd and then processed by udev rules – /etc/udev/rules.d/99-lustre.rules – SUBSYSTEM=="lustre", ACTION=="change", ENV{PARAM}=="?*", RUN+="/usr/sbin/lctl set_param '$env{PARAM}=$env{SETTING}’” • You can create your own udev rule – http://reactivated.net/writing_udev_rules.html – /lib/udev/rules.d/* for examples – Add udev_log="debug” to /etc/udev.conf if you have problems • Using systemd for long task.
    [Show full text]
  • Reducing Power Consumption in Mobile Devices by Using a Kernel
    IEEE TRANSACTIONS ON MOBILE COMPUTING, VOL. Z, NO. B, AUGUST 2017 1 Reducing Event Latency and Power Consumption in Mobile Devices by Using a Kernel-Level Display Server Stephen Marz, Member, IEEE and Brad Vander Zanden and Wei Gao, Member, IEEE E-mail: [email protected], [email protected], [email protected] Abstract—Mobile devices differ from desktop computers in that they have a limited power source, a battery, and they tend to spend more CPU time on the graphical user interface (GUI). These two facts force us to consider different software approaches in the mobile device kernel that can conserve battery life and reduce latency, which is the duration of time between the inception of an event and the reaction to the event. One area to consider is a software package called the display server. The display server is middleware that handles all GUI activities between an application and the operating system, such as event handling and drawing to the screen. In both desktop and mobile devices, the display server is located in the application layer. However, the kernel layer contains most of the information needed for handling events and drawing graphics, which forces the application-level display server to make a series of system calls in order to coordinate events and to draw graphics. These calls interrupt the CPU which can increase both latency and power consumption, and also require the kernel to maintain event queues that duplicate event queues in the display server. A further drawback of placing the display server in the application layer is that the display server contains most of the information required to efficiently schedule the application and this information is not communicated to existing kernels, meaning that GUI-oriented applications are scheduled less efficiently than they might be, which further increases power consumption.
    [Show full text]
  • I.MX Linux® Reference Manual
    i.MX Linux® Reference Manual Document Number: IMXLXRM Rev. 1, 01/2017 i.MX Linux® Reference Manual, Rev. 1, 01/2017 2 NXP Semiconductors Contents Section number Title Page Chapter 1 About this Book 1.1 Audience....................................................................................................................................................................... 27 1.1.1 Conventions................................................................................................................................................... 27 1.1.2 Definitions, Acronyms, and Abbreviations....................................................................................................27 Chapter 2 Introduction 2.1 Overview.......................................................................................................................................................................31 2.1.1 Software Base................................................................................................................................................ 31 2.1.2 Features.......................................................................................................................................................... 31 Chapter 3 Machine-Specific Layer (MSL) 3.1 Introduction...................................................................................................................................................................37 3.2 Interrupts (Operation)..................................................................................................................................................
    [Show full text]
  • Networking with Wicked in SUSE® Linux Enterprise 12
    Networking with Wicked in SUSE® Linux Enterprise 12 Something Wicked This Way Comes Guide www.suse.com Solution Guide Server Server Solution Guide Networking with Wicked in SUSE Linux Enterprise 12 Wicked QuickStart Guide Abstract: Introduced with SUSE® Linux Enterprise 12, Wicked is the new network management tool for Linux, largely replacing the sysconfig package to manage the ever-more-complicated network configurations. Wicked provides network configuration as a service, enabling you to change your configuration dynamically. This paper covers the basics of Wicked with an emphasis on Recently, new technologies have accelerated the trend toward providing correlations between how things were done previ- complexity. Virtualization requires on-demand provisioning of ously and how they need to be done now. resources, including networks. Converged networks that mix data and storage traffic on a shared link require a tighter integra- Introduction tion between stacks that were previously mostly independent. When S.u.S.E.1 first introduced its Linux distribution, network- ing requirements were relatively simple and static. Over time Today, more than 20 years after the first SUSE distribution, net- networking evolved to become far more complex and dynamic. work configurations are very difficult to manage properly, let For example, automatic address configuration protocols such as alone easily (see Figure 1). DHCP or IPv6 auto-configuration entered the picture along with a plethora of new classes of network devices. Modern Network Landscape While this evolution was happening, the concepts behind man- aging a Linux system’s network configuration didn’t change much. The basic idea of storing a configuration in some files and applying it at system boot up using a collection of scripts and system programs was pretty much the same.
    [Show full text]
  • O'reilly Linux Kernel in a Nutshell.Pdf
    ,title.4229 Page i Friday, December 1, 2006 9:52 AM LINUX KERNEL IN A NUTSHELL ,title.4229 Page ii Friday, December 1, 2006 9:52 AM Other Linux resources from O’Reilly Related titles Building Embedded Linux Running Linux Systems Understanding Linux Linux Device Drivers Network Internals Linux in a Nutshell Understanding the Linux Linux Pocket Guide Kernel Linux Books linux.oreilly.com is a complete catalog of O’Reilly’s Resource Center books on Linux and Unix and related technologies, in- cluding sample chapters and code examples. Conferences O’Reilly brings diverse innovators together to nurture the ideas that spark revolutionary industries. We spe- cialize in documenting the latest tools and systems, translating the innovator’s knowledge into useful skills for those in the trenches. Visit conferences.oreilly.com for our upcoming events. Safari Bookshelf (safari.oreilly.com) is the premier on- line reference library for programmers and IT professionals. Conduct searches across more than 1,000 books. Subscribers can zero in on answers to time-critical questions in a matter of seconds. Read the books on your Bookshelf from cover to cover or sim- ply flip to the page you need. Try it today for free. ,title.4229 Page iii Friday, December 1, 2006 9:52 AM LINUX KERNEL IN A NUTSHELL Greg Kroah-Hartman Beijing • Cambridge • Farnham • Köln • Paris • Sebastopol • Taipei • Tokyo ,LKNSTOC.fm.8428 Page v Friday, December 1, 2006 9:55 AM Chapter 1 Table of Contents Preface . ix Part I. Building the Kernel 1. Introduction . 3 Using This Book 4 2. Requirements for Building and Using the Kernel .
    [Show full text]
  • Communicating Between the Kernel and User-Space in Linux Using Netlink Sockets
    SOFTWARE—PRACTICE AND EXPERIENCE Softw. Pract. Exper. 2010; 00:1–7 Prepared using speauth.cls [Version: 2002/09/23 v2.2] Communicating between the kernel and user-space in Linux using Netlink sockets Pablo Neira Ayuso∗,∗1, Rafael M. Gasca1 and Laurent Lefevre2 1 QUIVIR Research Group, Departament of Computer Languages and Systems, University of Seville, Spain. 2 RESO/LIP team, INRIA, University of Lyon, France. SUMMARY When developing Linux kernel features, it is a good practise to expose the necessary details to user-space to enable extensibility. This allows the development of new features and sophisticated configurations from user-space. Commonly, software developers have to face the task of looking for a good way to communicate between kernel and user-space in Linux. This tutorial introduces you to Netlink sockets, a flexible and extensible messaging system that provides communication between kernel and user-space. In this tutorial, we provide fundamental guidelines for practitioners who wish to develop Netlink-based interfaces. key words: kernel interfaces, netlink, linux 1. INTRODUCTION Portable open-source operating systems like Linux [1] provide a good environment to develop applications for the real-world since they can be used in very different platforms: from very small embedded devices, like smartphones and PDAs, to standalone computers and large scale clusters. Moreover, the availability of the source code also allows its study and modification, this renders Linux useful for both the industry and the academia. The core of Linux, like many modern operating systems, follows a monolithic † design for performance reasons. The main bricks that compose the operating system are implemented ∗Correspondence to: Pablo Neira Ayuso, ETS Ingenieria Informatica, Department of Computer Languages and Systems.
    [Show full text]
  • Beancounters
    Resource Management: Beancounters Pavel Emelianov Denis Lunev Kirill Korotaev [email protected] [email protected] [email protected] July 31, 2007 Abstract The paper outlines various means of resource management available in the Linux kernel, such as per-process limits (the setrlimit(2) interface), shows their shortcomings, and illustrares the need for another resource control mechanism: beancounters. Beancounters are a set of per-process group parameters (proposed and implemented by Alan Cox and Andrey Savochkin and further developed for OpenVZ) which can be used with or without containers. Beancounters history, architecture, goals, efficiency, and some in-depth implementation details are given. 1 Current state of resource Again, most of these resource limits apply to a management in the Linux single process, which means, for example, that all the memory may be consumed by a single user run- kernel ning the appropriate number of processes. Setting the limits in such a way as to have the value of Currently the Linux kernel has only one way to multiplying the per-process limit by the number of control resource consumption of running processes processes staying below the available values is im- – it is UNIX-like resource limits (rlimits). practical. Rlimits set upper bounds for some resource us- age parameters. Most of these limits apply to a single process, except for the limit for the number 2 Beancounters of processes, which applies to a user. For some time now Linus has been accepting The main goal of these limits is to protect pro- patches adding the so-called namespaces into the cesses from an accidental misbehavior (like infi- kernel.
    [Show full text]
  • Postmodern Strace Dmitry Levin
    Postmodern strace Dmitry Levin Brussels, 2020 Traditional strace [1/30] Printing instruction pointer and timestamps print instruction pointer: -i option print timestamps: -r, -t, -tt, -ttt, and -T options Size and format of strings string size: -s option string format: -x and -xx options Verbosity of syscall decoding abbreviate output: -e abbrev=set, -v option dereference structures: -e verbose=set print raw undecoded syscalls: -e raw=set Traditional strace [2/30] Printing signals print signals: -e signal=set Dumping dump the data read from the specified descriptors: -e read=set dump the data written to the specified descriptors: -e write=set Redirecting output to files or pipelines write the trace to a file or pipeline: -o filename option write traces of processes to separate files: -ff -o filename Traditional strace [3/30] System call filtering trace only the specified set of system calls: -e trace=set System call statistics count time, calls, and errors for each system call: -c option sort the histogram printed by the -c option: -S sortby option Tracing control attach to existing processes: -p pid option trace child processes: -f option Modern strace [4/30] Tracing output format pathnames accessed by name or descriptor: -y option network protocol associated with descriptors: -yy option stack of function calls: -k option System call filtering pathnames accessed by name or descriptor: -P option regular expressions: -e trace=/regexp optional specifications: -e trace=?spec new syscall classes: %stat, %lstat, %fstat, %statfs, %fstatfs, %%stat, %%statfs
    [Show full text]
  • Compression and Replication of Device Files Using Deduplication Technique
    Compression and Replication of Device Files using Deduplication Technique {tag} {/tag} International Journal of Computer Applications © 2012 by IJCA Journal Volume 45 - Number 1 Year of Publication: 2012 Authors: Bharati Siddhant Agarwal Abhishek Somani Rukmi Patel Ankita Shingvi 10.5120/6744-8847 {bibtex}pxc3878847.bib{/bibtex} Abstract One of the biggest challenges to the data storage community is how to effectively store data without taking the exact same data and storing again and again in different locations on the back end servers. In this paper, we outline the performance achievements that can be achieved by exploiting block level data deduplication in File Systems. We have achieved replication of data by storing a deduplicated copy of the original data on the backup partition, which can enable us to access or restore data in case of a system crash. References - Request-based Device-mapper multipath and Dynamic load balancing by kiyoshi Ueda, Jun'ichi Nomura and Mike Christie. 1 / 2 Compression and Replication of Device Files using Deduplication Technique - A Device Mapper based Encryption Layer for TransCrypt by Sainath Vella. - Edward Goggin, Linux Multipathing Proceedings of the Linux Symposium, 2005. - Jens Axboe, Notes on the Generic Block Layer Rewrite in Linux 2. 5, Documentation/block/ biodoc. txt, 2007. - Netlink - Communication between kernel and userspace (PF NETLINK). Manual Page Netlink. - Satyam Sharma, Rajat Moona, and Dheeraj Sanghi. TransCrypt: A Secure and Transparent Encrypting File System for Enterprises. - Mike Anderson, SCSI Mid-Level Multipath, Proceedings of the Linux Symposium, 2003. - Daniel Pierre Bovet and Marco Cesati. Understanding the Linux Kernel. O'Reilly& Associates, Inc.
    [Show full text]
  • Daemon Management Under Systemd ZBIGNIEWSYSADMIN JĘDRZEJEWSKI-SZMEK and JÓHANN B
    Daemon Management Under Systemd ZBIGNIEWSYSADMIN JĘDRZEJEWSKI-SZMEK AND JÓHANN B. GUÐMUNDSSON Zbigniew Jędrzejewski-Szmek he systemd project is the basic user-space building block used to works in a mixed experimental- construct a modern Linux OS. The main daemon, systemd, is the first computational neuroscience lab process started by the kernel, and it brings up the system and acts as and writes stochastic simulators T and programs for the analysis a service manager. This article shows how to start a daemon under systemd, of experimental data. In his free time he works describes the supervision and management capabilities that systemd pro- on systemd and the Fedora Linux distribution. vides, and shows how they can be applied to turn a simple application into [email protected] a robust and secure daemon. It is a common misconception that systemd is somehow limited to desktop distributions. This is hardly true; similarly to Jóhann B. Guðmundsson, the Linux kernel, systemd supports and is used on servers and desktops, but Penguin Farmer, IT Fireman, Archer, Enduro Rider, Viking- it is also in the cloud and extends all the way down to embedded devices. In Reenactor, and general general it tries to be as portable as the kernel. It is now the default on new insignificant being in an installations in Debian, Ubuntu, Fedora/RHEL/CentOS, OpenSUSE/SUSE, insignificant world, living in the middle of the Arch, Tizen, and various derivatives. North Atlantic on an erupting rock on top of the world who has done a thing or two in Systemd refers both to the system manager and to the project as a whole.
    [Show full text]
  • Interaction Between the User and Kernel Space in Linux
    1 Interaction Between the User and Kernel Space in Linux Kai Lüke, Technische Universität Berlin F Abstract—System calls based on context switches from user to kernel supported in POSIX as well as System V style. A very space are the established concept for interaction in operating systems. common principle for IPC are sockets, and pipes can be seen On top of them the Linux kernel offers various paradigms for commu- as their most simple case. Besides the popular IP family with nication and management of resources and tasks. The principles and TCP/UDP, the local Unix domain sockets play a big role basic workings of system calls, interrupts, virtual system calls, special for many applications, while Netlink sockets are specific purpose virtual filesystems, process signals, shared memory, pipes, Unix or IP sockets and other IPC methods like the POSIX or System V to the Linux kernel and are not often found in user space message queue and Netlink are are explained and related to each other applications due to portability to other operating systems. in their differences. Because Linux is not a puristic project but home for There have been attempts to bring the D-BUS IPC into many different concepts, only a mere overview is presented here with the kernel with a Netlink implementation, kdbus and then focus on system calls. Bus1, but this will not be covered here. Also comparative studies with other operating systems are out of scope. 1 INTRODUCTION 2 KERNEL AND USER SPACE ERNELS in the Unix family are normally not initiating K any actions with outside effects, but rely on requests The processes have virtualized access to the memory as well from user space to perform these actions.
    [Show full text]
  • Instant OS Updates Via Userspace Checkpoint-And
    Instant OS Updates via Userspace Checkpoint-and-Restart Sanidhya Kashyap, Changwoo Min, Byoungyoung Lee, and Taesoo Kim, Georgia Institute of Technology; Pavel Emelyanov, CRIU and Odin, Inc. https://www.usenix.org/conference/atc16/technical-sessions/presentation/kashyap This paper is included in the Proceedings of the 2016 USENIX Annual Technical Conference (USENIX ATC ’16). June 22–24, 2016 • Denver, CO, USA 978-1-931971-30-0 Open access to the Proceedings of the 2016 USENIX Annual Technical Conference (USENIX ATC ’16) is sponsored by USENIX. Instant OS Updates via Userspace Checkpoint-and-Restart Sanidhya Kashyap Changwoo Min Byoungyoung Lee Taesoo Kim Pavel Emelyanov† Georgia Institute of Technology †CRIU & Odin, Inc. # errors # lines Abstract 50 1000K 40 100K In recent years, operating systems have become increas- 10K 30 1K 20 ingly complex and thus more prone to security and per- 100 formance issues. Accordingly, system updates to address 10 10 these issues have become more frequently available and 0 1 increasingly important. To complete such updates, users 3.13.0-x 3.16.0-x 3.19.0-x May 2014 must reboot their systems, resulting in unavoidable down- build/diff errors #layout errors Jun 2015 time and further loss of the states of running applications. #static local errors #num lines++ We present KUP, a practical OS update mechanism that Figure 1: Limitation of dynamic kernel hot-patching using employs a userspace checkpoint-and-restart mechanism, kpatch. Only two successful updates (3.13.0.32 34 and → which uses an optimized data structure for checkpoint- 3.19.0.20 21) out of 23 Ubuntu kernel package releases.
    [Show full text]