Unifying Hardware and Software Benchmarking: a Resource-Agnostic Model

Total Page:16

File Type:pdf, Size:1020Kb

Unifying Hardware and Software Benchmarking: a Resource-Agnostic Model Unifying hardware and software benchmarking: a resource-agnostic model Davide Morelli University of Pisa Computer Science Department June 2015 Ph.D. thesis 2 Abstract Lilja (2005) states that \In the field of computer science and engineering there is surprisingly little agreement on how to measure something as fun- damental as the performance of a computer system.". The field lacks of the most fundamental element for sharing measures and results: an appropriate metric to express performance. Since the introduction of laptops and mobile devices, there has been a strong research focus towards the energy efficiency of hardware. Many papers, both from academia and industrial research labs, focus on methods and ideas to lower power consumption in order to lengthen the battery life of portable device components. Much less effort has been spent on defining the responsibility of software in the overall computational system energy consumption. Some attempts have been made to describe the energy behaviour of software, but none of them abstract from the physical machine where the measurements were taken. In our opinion this is a strong drawback because results can not be generalized. In this work we attempt to bridge the gap between characterization and prediction, of both hardware and software, of performance and energy, in a single unified model. We propose a model designed to be as simple as possible, generic enough to be abstract from the specific resource being described or predicted (applying to both time, memory and energy), but also concrete and practical, allowing useful and precise performance and energy predictions. The model applies to the broadest set of resource possible. We focus mainly on time and memory (hence bridging hardware benchmarking and classical algorithms time complexity), and energy consumption. To ensure a wide applicability of the model in real world scenario, the model is completely black-box, it does not require any information about the source code of the program, and only relies on external metrics, like completion time, energy consumption, or performance counters. Extending the benchmarking model, we define the notion of experimental computational complexity, as the characterization of how the resource usage 3 4 changes as the input size grows. Finally, we define a high-level energy model capable of characterizing the power consumption of computers and clusters, in terms of the usage of resources as defined by our benchmarking model. We tested our model in four experiments: Expressiveness we show the close relationship between energy and clas- sical theoretical complexity, also showing that our experimental com- putational complexity is expressive enough to capture interesting be- haviour of programs simply analysing their resource usage. Performance prediction we use the large database of performance mea- sures available on the CPU SPEC website to train our model and predict the performance of the CPU SPEC suite on randomly selected computers. Energy profiling we tested our model to characterize and predict the power usage of a cluster running OpenFOAM, changing the number of active nodes and cores. Scheduling on heterogeneous systems applying our performance pre- diction model to features of programs extracted at runtime, we predict the device where is most convenient to execute the programs, in an heterogeneous system. Acknowledgements I want to thank my supervisor Antonio Cisternino, for the constant insight, for ensuring I was following the right path in my research, and for his friend- ship. The advice from Andrea Canciani has been extremely important to assess the methodology in general, and to develop the energy model in particular. Leonardo Bartoloni has helped me understanding the probabilistic frame- work underlying a model based on regression analysis. The last experiment reported in this thesis has been developed with Gabriele Cocco, using his Ph.D. research we created a scheduling method- ology for heterogeneous platforms. I am grateful for the support and constant affect I received from my family: my parents, my grandparents, and my sister, who supported and encouraged me throughout this journey. I wish to thank my wife for sharing her life with me, constantly pushing me to pursue my dreams. 5 6 Contents I Introduction 19 1 Benchmarking: art or science 21 1.1 Introduction . 21 1.2 Benchmarking is pre-science . 23 2 Lack of a unified model 25 2.1 Time and energy . 26 2.2 Hardware and Software . 27 2.3 Computational energy models . 28 3 Contributions of this work 31 3.1 Contributions . 31 3.2 Plan of work . 32 4 State of the art: characterization and prediction of perfor- mance and energy 35 4.1 Hardware benchmarking . 36 4.1.1 Metrics . 36 4.1.2 Benchmarks . 39 4.2 Software characterization . 41 4.3 Energy characterization . 44 4.3.1 Power measurement approaches . 44 4.3.2 Energy consumption of virtual machines . 46 4.3.3 Energy consumption related to parallelism . 46 4.3.4 Energy consumption related to time complexity . 47 4.3.5 Energy management policies . 47 4.3.6 Coarse grain energy models . 47 4.4 Scheduling on heterogeneous architectures . 50 7 8 CONTENTS II Benchmarking Model 53 5 High level model 57 5.1 Definitions . 58 5.1.1 Program . 58 5.1.2 Computational Environment . 58 5.1.3 Measure of resource consumption . 59 5.1.4 Computational Pattern . 61 5.1.5 Solver . 61 5.2 The model . 62 5.2.1 Characterizing a target program . 63 5.2.2 Characterizing the target resource . 65 5.2.3 A unified HW and SW model . 66 5.2.4 A resource agnostic model . 67 5.3 Role of computational patterns . 68 5.4 Algebraic characterization of X ................. 71 5.4.1 X needs to be full rank . 71 5.4.2 Norm . 71 5.4.3 Cosine similarity . 73 5.5 Conclusion . 75 6 Solvers 77 6.1 Simplex . 77 6.1.1 The simplex algorithm . 78 6.1.2 Geometric interpretation . 79 6.1.3 Limits . 82 6.2 Linear regression . 84 6.2.1 Example . 84 6.2.2 Linear regression . 86 6.2.3 Algebraic characterization of the solver . 87 6.2.4 Hardware and Software intrinsic relationship . 90 6.2.5 Assumptions and limits . 90 6.2.6 Residuals analysis . 91 6.2.7 Algebraic characterization of computational patterns . 92 6.3 Non negative matrix factorization . 95 6.3.1 The surrogate . 96 6.3.2 The predictor . 97 6.3.3 Hardware and Software intrinsic relationship . 98 6.3.4 Limits of NMF . 99 6.4 Conclusion and future work . 100 CONTENTS 9 6.4.1 Bayesian regression . 100 6.4.2 Support Vector Regression . 102 7 Experimental complexity of software 105 7.1 Characterization through the surrogate . 105 7.2 Definition of experimental complexity . 106 7.3 Relation with computational complexity . 109 7.4 Bottlenecks . 109 7.4.1 Grace area . 110 7.5 Compositionality of Surrogates and Experimental complexity 111 7.5.1 Linear composition of surrogates and experimental com- plexity ........................... 111 7.5.2 Function composition . 112 7.6 A single number is not enough . 113 7.7 Conclusions . 117 8 Computational energy model 119 8.1 The energy model . 120 8.1.1 General form . 120 8.1.2 Fixed number of active machines and limited consid- ered resources . 122 8.1.3 Automatic characterization of hardware . 124 8.2 Discussion . 125 8.3 Conclusions . 129 9 Conclusions 131 III Experimental validation 133 10 Validating the expressiveness of the model and experimental complexity using micro-benchmarks 137 10.1 Expressiveness of the model . 138 10.1.1 Using the simplex solver . 138 10.1.2 Using the linear regression and NMF solvers . 140 10.2 Experimental computational complexity . 145 10.2.1 The experimental setup . 145 10.2.2 ξ is compositional . 149 10.3 Energy and experimental computational complexity: sorting algorithms . 149 10 CONTENTS 10.3.1 Experiment setting . 152 10.3.2 Experimental results . 152 10.4 Micro-architecture independence: limits of a single dimension 154 10.4.1 Different architectures . 155 10.4.2 Using all the measures in the same model, only 1 pro- gram as basis . 158 10.4.3 Using all the measures in the same model, with a larger basis . 159 10.4.4 Using the target program as basis . 160 10.4.5 Discussion . 160 10.5 Conclusions . 162 11 Validating performance prediction on multiple architectures using CPUSPEC 163 11.1 CPUSPEC . 163 11.1.1 Bias error . 164 11.2 Predicting completion time using the linear regression solver . 164 11.2.1 Completion time prediction accuracy . 167 11.2.2 Fitting residuals, bias, and prediction error . 169 11.2.3 Discussion . 172 11.3 Conclusions . 172 12 Validating the energy model for concurrent parallel tasks using OpenFOAM 177 12.1 Predict peak power . 177 12.1.1 Experiment setting . 178 12.1.2 Modelling power using Pinfr+m and P∆c, without sep- arating Pinfr and Pm ................... 180 12.1.3 Modelling power using Pinfr,P∆c, and Pm ....... 182 12.2 Modelling energy consumption of concurrent programs . 184 12.3 Conclusions . 190 13 Combining static analysis with the prediction model: best device scheduling in heterogeneous environments 191 13.1 Introduction . 192 13.2 Code analysis and feature extraction . 193 13.3 Prediction model . 194 13.4 Experimental validation . 197 13.4.1 Experiment setup . 197 13.4.2 Predicting the completion time . 201 CONTENTS 11 13.4.3 Best-device prediction . 204 13.4.4 Interpretation of the regression coefficients . 205 13.4.5 Limits . 207 13.5 Conclusions . 208 IV Conclusions 209 14 Contributions 213 15 Future work 217 12 CONTENTS List of Figures 6.1 Convex polytope of solutions in the resource consumption vec- tor space .............................
Recommended publications
  • Personal Computing, the Notebook Battery Crisis, and Postindustrial
    1 1 <running head>Eisler | Exploding the Black Box 2 <title>Exploding the Black Box 3 <subtitle>Personal Computing, the Notebook Battery Crisis, and Postindustrial 4 Systems Thinking 5 Matthew N. Eisler 6 Matthew N. Eisler studies the relationship between ideology, material practices, and 7 the social relations of contemporary science and engineering at the intersection of energy, 8 environmental, and industrial policy. His first book, Overpotential: Fuel Cells, Futurism, 9 and the Making of a Power Panacea was published by Rutgers University Press in 2012. 10 He has been affiliated with Western University, the Center for Nanotechnology in Society 11 at the University of California at Santa Barbara, the Chemical Heritage Foundation, and 12 the Department of Engineering and Society at the University of Virginia. He is currently a 13 Visiting Assistant Professor in the Department of Integrated Science and Technology at 14 James Madison University. He thanks Jack K. Brown, W. Bernard Carlson, Paul E. 15 Ceruzzi, Michael D. Gordin, Barbara Hahn, Cyrus C.M. Mody, Hannah S. Rogers, and 16 three anonymous referees for their incisive and constructive criticism of earlier drafts of 17 this article. 18 Abstract: 19 Historians of science and technology have generally ignored the role of power 20 sources in the development of consumer electronics. In this they have followed the 21 predilections of historical actors. Research, development, and manufacturing of 22 batteries has historically occurred at a social and intellectual distance from the research, 23 development, and manufacturing of the devices they power. Nevertheless, power source 24 technoscience should properly be understood as an allied yet estranged field of 2 1 electronics.
    [Show full text]
  • Metrics for Performance Advertisement
    Metrics for Performance Advertisement Vojtěch Horký Peter Libič Petr Tůma 2010 – 2021 This work is licensed under a “CC BY-NC-SA 3.0” license. Created to support the Charles University Performance Evaluation lecture. See http://d3s.mff.cuni.cz/teaching/performance-evaluation for details. Contents 1 Overview 1 2 Operation Frequency Related Metrics 2 3 Operation Duration Related Metrics 3 4 Benchmark Workloads 4 1 Overview Performance Advertisement Measuring for the purpose of publishing performance information. Requirements: – Well defined meaning. – Simple to understand. – Difficult to game. Pitfalls: – Publication makes results subject to pressure. – Often too simple to convey meaningful information. – Performance in contemporary computer systems is never simple. Speed Related Metrics Responsiveness Metrics – Time (Task Response Time) – How long does it take to finish the task ? Productivity Metrics – Rate (Task Throughput) – How many tasks can the system complete per time unit ? Utilization Metrics – Resource Use (Utilization) – How much is the system loaded when working on a task ? – Share of time a resource is busy or over given load level. – Helps identify bottlenecks (most utilized resources in the system). 1 Metric for Webmail Performance What metric would you choose to characterize performance of a web mail site ? – User oriented metric would be end-to-end operation time. – Server oriented metric would be request processing time. – How about metrics in between ? – And would you include mail delivery time ? How is throughput related to latency ? How is utilization defined for various resources ? 2 Operation Frequency Related Metrics Clock Rate Clock rate (frequency) of the component (CPU, bus, memory) in MHz. Most often we talk about CPU frequency.
    [Show full text]
  • The Nutanix Design Guide
    The FIRST EDITION Nutanix Design Guide Edited by AngeloLuciani,VCP By Nutanix, RenévandenBedem, incollaborationwith NPX,VCDX4,DECM-EA RoundTowerTechnologies Table of Contents 1 Foreword 3 2 Contributors 5 3 Acknowledgements 7 4 Using This Book 7 5 Introduction To Nutanix 9 6 Why Nutanix? 17 7 The Nutanix Eco-System 37 8 Certification & Training 43 9 Design Methodology 49 & The NPX Program 10 Channel Charter 53 11 Mission-Critical Applications 57 12 SAP On Nutanix 69 13 Hardware Platforms 81 14 Sizer & Collector 93 15 IBM Power Systems 97 16 Remote Office, Branch Office 103 17 Xi Frame & EUC 113 Table of Contents 18 Xi IoT 121 19 Xi Leap, Data Protection, 133 DR & Metro Availability 20 Cloud Management 145 & Automation: Calm, Xi Beam & Xi Epoch 21 Era 161 22 Karbon 165 23 Acropolis Security & Flow 169 24 Files 193 25 Volumes 199 26 Buckets 203 27 Prism 209 28 Life Cycle Manager 213 29 AHV 217 30 Move 225 31 X-Ray 229 32 Foundation 241 33 Data Center Facilities 245 34 People & Process 251 35 Risk Management 255 The Nutanix Design Guide 1 Foreword Iamhonoredtowritethisforewordfor‘TheNutanixDesignGuide’. Wehavealwaysbelievedthatcloudwillbemorethanjustarented modelforenterprises.Computingwithintheenterpriseisnuanced, asittriestobalancethefreedomandfriction-freeaccessofthe publiccloudwiththesecurityandcontroloftheprivatecloud.The privateclouditselfisspreadbetweencoredatacenters,remoteand branchoffices,andedgeoperations.Thetrifectaofthe3laws–(a) LawsoftheLand(dataandapplicationsovereignty),(b)Lawsof Physics(dataandmachinegravity),and(c)LawsofEconomics
    [Show full text]
  • “Freedom” Koan-Sin Tan [email protected] OSDC.Tw, Taipei Apr 11Th, 2014
    Understanding Android Benchmarks “freedom” koan-sin tan [email protected] OSDC.tw, Taipei Apr 11th, 2014 1 disclaimers • many of the materials used in this slide deck are from the Internet and textbooks, e.g., many of the following materials are from “Computer Architecture: A Quantitative Approach,” 1st ~ 5th ed • opinions expressed here are my personal one, don’t reflect my employer’s view 2 who am i • did some networking and security research before • working for a SoC company, recently on • big.LITTLE scheduling and related stuff • parallel construct evaluation • run benchmarking from time to time • for improving performance of our products, and • know what our colleagues' progress 3 • Focusing on CPU and memory parts of benchmarks • let’s ignore graphics (2d, 3d), storage I/O, etc. 4 Blackbox ! • google image search “benchmark”, you can find many of them are Android-related benchmarks • Similar to recently Cross-Strait Trade in Services Agreement (TiSA), most benchmarks on Android platform are kinda blackbox 5 Is Apple A7 good? • When Apple released the new iPhone 5s, you saw many technical blog showed some benchmarks for reviews they came up • commonly used ones: • GeekBench • JavaScript benchmarks • Some graphics benchmarks • Why? Are they right ones? etc. e.g., http://www.anandtech.com/show/7335/the-iphone-5s-review 6 open blackbox 7 Android Benchmarks 8 http:// www.anandtech.com /show/7384/state-of- cheating-in-android- benchmarks No, not improvement in this way 9 Assuming there is not cheating, what we we can do? Outline • Performance benchmark review • Some Android benchmarks • What we did and what still can be done • Future 11 To quote what Prof.
    [Show full text]
  • CS61C : Machine Structures
    inst.eecs.berkeley.edu/~cs61c/su05 Review CS61C : Machine Structures • Benchmarks Lecture #29: Intel & Summary • Attempt to predict performance • Updated every few years • Measure everything from simulation of desktop graphics programs to battery life • Megahertz Myth • MHz ≠ performance, it’s just one factor 2005-08-10 Andy Carle CS 61C L29 Intel & Review (1) A Carle, Summer 2005 © UCB CS 61C L29 Intel & Review (2) A Carle, Summer 2005 © UCB MIPS is example of RISC MIPS vs. 80386 • RISC = Reduced Instruction Set Computer • Address: 32-bit • 32-bit • Term coined at Berkeley, ideas pioneered • Page size: 4KB • 4KB by IBM, Berkeley, Stanford • Data aligned • Data unaligned • RISC characteristics: • Destination reg: Left • Right • Load-store architecture •add $rd,$rs1,$rs2 •add %rs1,%rs2,%rd • Fixed-length instructions (typically 32 bits) • Regs: $0, $1, ..., $31 • %r0, %r1, ..., %r7 • Three-address architecture • Reg = 0: $0 • (n.a.) • RISC examples: MIPS, SPARC, • Return address: $31 • (n.a.) IBM/Motorola PowerPC, Compaq Alpha, ARM, SH4, HP-PA, ... CS 61C L29 Intel & Review (3) A Carle, Summer 2005 © UCB CS 61C L29 Intel & Review (4) A Carle, Summer 2005 © UCB MIPS vs. Intel 80x86 MIPS vs. Intel 80x86 • MIPS: “load-store architecture” • MIPS: “Three-address architecture” • Only Load/Store access memory; rest • Arithmetic-logic specify all 3 operands operations register-register; e.g., add $s0,$s1,$s2 # s0=s1+s2 lw $t0, 12($gp) add $s0,$s0,$t0 # s0=s0+Mem[12+gp] • Benefit: fewer instructions ⇒ performance • Benefit: simpler hardware ⇒ easier
    [Show full text]
  • Red Hat Enterprise Virtualization Performance: Specvirt™ Benchmark
    RED HAT ENTERPRISE VIRTUALIZATION PERFORMANCE: SPECVIRT™ BENCHMARK DATASHEET AT A GLANCE OVERVIEW ® • The performance of Red Hat Red Hat Enterprise Virtualization (RHEV) is the strategic virtualization alternative for organizations Enterprise Virtualization can looking for better total cost of ownership, faster return on investment, accelerated break-even, and be compared to other freedom from vendor lock-in. virtualization platforms RHEV consists of both a hypervisor technology and an enterprise virtualization manager. The RHEV using the SPECvirt_sc2010 hypervisor, based on the Red Hat Enterprise Linux® kernel and the Kernel-based Virtual Machine benchmark. (KVM) hypervisor technology, offers the high performance and scalability that Red Hat is known for. • SPECvirt_sc2010 measures Hypervisor performance is a key factor in deciding which virtualization platform to implement, and hypervisor performance using Red Hat Enterprise Virtualization performance, as measured by the SPECvirt_sc2010® benchmark, realistic workloads. leads the industry both in terms of highest overall performance and highest number of performant As of March 1, 2013, RHEV virtual machines on a single server. leads VMware in performance on comparable WHAT IS SPECVIRT_SC2010®? servers defined by CPU count. SPECvirt_sc2010 is the first vendor-neutral benchmark designed to measure the performance of • The RHEV subscription model datacenter servers that are used for server virtualization. The benchmark was developed by the provides high performance non-profit Standard Performance Evaluation Corporation (SPEC) virtualization subcommittee, at a lower cost-per-SPECvirt_ whose members and contributors include AMD, Dell, Fujitsu, HP, IBM, Intel, Oracle, Red Hat, Unisys sc2010 than VMware. and VMware. SPECvirt_sc2010 uses realistic workloads and SPEC’s proven performance measurement method- ologies to enable vendors, users and researchers to compare system performance across multiple hardware, virtualization platforms, and applications.
    [Show full text]
  • On Benchmarking Intrusion Detection Systems in Virtualized Environments
    Technical Report: SPEC-RG-2013-002 Version: 1.0 On Benchmarking Intrusion Detection Systems in Virtualized Environments SPEC RG IDS Benchmarking Working Group Aleksandar Milenkoski Samuel Kounev Alberto Avritzer Institute for Program Structures and Data Institute for Program Structures and Data Siemens Corporate Research Organization Organization Princeton, NJ USA Karlsruhe Institute of Technology Karlsruhe Institute of Technology [email protected] Karlsruhe, Germany Karlsruhe, Germany [email protected] [email protected] Nuno Antunes Marco Vieira CISUC, Department of Informatics CISUC, Department of Informatics Engineering Engineering University of Coimbra University of Coimbra Coimbra, Portugal Coimbra, Portugal [email protected] [email protected] arXiv:1410.1160v1 [cs.CR] 5 Oct 2014 ® Research ℠ June 26, 2013 research.spec.org www.spec.org Contents 1 Introduction.......................................1 2 Intrusion Detection in Virtualized Environments . .2 2.1 VMM-Based Intrusion Detection Systems . .2 2.2 Intrusion Detection Techniques . .4 Misuse-based Intrusion Detection . .4 Anomaly-based Intrusion Detection . .5 3 Requirements and Challenges for Benchmarking VMM-based IDSes . .7 3.1 Workloads . .7 Benign Workloads . .7 Malicious Workloads . 10 3.2 Metrics ..................................... 14 4 Conclusion . 17 References . 18 i Executive Summary Modern intrusion detection systems (IDSes) for virtualized environments are deployed in the virtualization layer with components inside the virtual machine monitor (VMM) and the trusted host virtual machine (VM). Such IDSes can monitor at the same time the network and host activities of all guest VMs running on top of a VMM being isolated from malicious users of these VMs. We refer to IDSes for virtualized environments as VMM-based IDSes. In this work, we analyze state-of-the-art intrusion detection techniques applied in virtualized environments and architectures of VMM-based IDSes.
    [Show full text]
  • IADIS Conference Template
    BENCHMARKING OF BARE METAL VIRTUALIZATION PLATFORMS ON COMMODITY HARDWARE Duarte Pousa*, José Rufino*† *Polytechnic Institute of Bragança, 5300-253 Bragança, Portugal †Laboratory of Instrumentation and Experimental Particle Physics, University of Minho, 4710-057 Braga, Portugal ([email protected], [email protected]) ABSTRACT In recent years, System Virtualization became a fundamental IT tool, whether it is type-2/hosted virtualization, mostly exploited by end-users in their personal computers, or type-1/bare metal, well established in IT departments and thoroughly used in modern datacenters as the very foundation of cloud computing. Though bare metal virtualization is meant to be deployed on server-grade hardware (for performance, stability and reliability reasons), properly configured desktop-class systems are often used as virtualization “servers”, due to their attractive performance/cost ratio. This paper presents the results of a study conducted on such systems, about the performance of Windows 10 and Ubuntu Server 16.04 guests, when deployed in what we believe are the type-1 platforms most in use today: VMware ESXi, Citrix XenServer, Microsoft Hyper-V, and KVM-based (represented by oVirt and Proxmox). Performance is measured using three synthetic benchmarks: PassMark for Windows, UnixBench for Ubuntu Server, and the cross-platform Flexible I/O Tester. The benchmarks results may be used to choose the most adequate type-1 platform (performance-wise), depending on guest OS, its performance requisites (CPU-bound, IO-bound, or balanced) and its storage type (local/remote) used. KEYWORDS Bare Metal Virtualization, Synthetic Benchmarking, Performance Assessment, Commodity Hardware 1. INTRODUCTION System Virtualization allows a physical host to run (simultaneously) many guest operating systems in self- contained software-defined virtual machines, sharing the host hardware, under control of a hypervisor.
    [Show full text]
  • TPC-V: a Benchmark for Evaluating the Performance of Database Applications in Virtual Environments Priya Sethuraman and H
    TPC-V: A Benchmark for Evaluating the Performance of Database Applications in Virtual Environments Priya Sethuraman and H. Reza Taheri, VMWare, Inc. {psethuraman, rtaheri}@vmware.com TPCTC 2010 Singapore © 2010 VMware Inc. All rights reserved Agenda/Topics Introduction to virtualization Existing benchmarks Genesis of TPC-V But what is TPC-E??? TPC-V design considerations Set architectures, variability, and elasticity Benchmark development status Answers to some common questions 2 TPC TC 2010 What is a Virtual Machine? A (VM) is a software computer that, like a physical computer, runs an operating system and applications. An operating system installed on a virtual machine is called a guest operating system. Virtual machines run on host servers. The same server can run many virtual machines. Every VM runs in an isolated environment. Started out with IBM VM in the 60s Also on Sun Solaris, HP Itanium, IBM Power/AIX, others A new wave started in the late 90s on X86 • Initially, enthusiasts ran Windows and Linux VMs on their PCs Traditional Architecture Virtual Architecture 3 TPC TC 2010 Why virtualize a server ? Server consolidation • The vast majority of server are grossly underutilized • Reduces both CapEx and OpEx Migration of VMs (both storage and CPU/memory) • Enables live load balancing • Facilitates maintenance High availability • Allows a small number of generic servers to back up all servers Fault tolerance • Lock-step execution of two VMs Cloud computing! Utility computing was finally enabled by • Ability to consolidate
    [Show full text]
  • KVM Virtualization Roadmap and Technology Update
    KVM Virtualization Roadmap and Technology Update Karen Noel Bhavna Sarathy Senior Eng Manager Senior Product Manager Red Hat, Inc. Red Hat, Inc. June 13, 2013 Why we believe KVM is the best virtualization platform Performance Lower Cost Cross Platform KVM holds the Top 6/11 customers report up to Support and certification for virtual machine consolidation 70% savings by using leading x86_64 operating systems scores on SPECvirt (1) KVM (2) including RHEL and Microsoft Windows (4) Security Cloud & Virtualization Management Certification (3) EAL4+ Red Hat Open Stack for Cloud plus SE Linux enabling Virtualization and Red Hat Mandatory Access Control Enterprise Virtualization for data- between virtual machines center Virtualization (1) Source: SpecVirt_sc2010 results: http://www.spec.org/virt_sc2010/results/specvirt_sc2010_perf.html (2) Source: Case study on Canary Islands Government migration from VMware to RHEV: http://www.redhat.com/resourcelibrary/case-studies/canary-islands-government-migrates-telecommunications-platform-from- vmware-to-red-hat (3) Source: http://www.redhat.com/solutions/industry/government/certifications.html (4) Source: http://www.redhat.com/resourcelibrary/articles/enterprise-linux-virtualization-support KVM hypervisor in multiple Red Hat products KVM is the foundation Virtualization technology in multiple Red Hat products Red Hat Enterprise Virtualization – Hypervisor Derived from Red Hat Enterprise Linux SMALL FORM FACTOR, SCALABLE, ● RHEV Hypervisor HIGH PERFORMANCE ● Prebuilt binary (ISO) with 300+ packages derived
    [Show full text]
  • CS61C : Machine Structures
    inst.eecs.berkeley.edu/~cs61c/su05 Review CS61C : Machine Structures • Benchmarks Lecture #29: Intel & Summary • Attempt to predict performance • Updated every few years • Measure everything from simulation of desktop graphics programs to battery life • Megahertz Myth • MHz ≠ performance, it’s just one factor 2005-08-10 Andy Carle CS 61C L29 Intel & Review (1) A Carle, Summer 2005 © UCB CS 61C L29 Intel & Review (2) A Carle, Summer 2005 © UCB MIPS is example of RISC MIPS vs. 80386 • RISC = Reduced Instruction Set Computer • Address: 32-bit • 32-bit • Term coined at Berkeley, ideas pioneered • Page size: 4KB • 4KB by IBM, Berkeley, Stanford • Data aligned • Data unaligned • RISC characteristics: • Destination reg: Left • Right • Load-store architecture •add $rd,$rs1,$rs2 •add %rs1,%rs2,%rd • Fixed-length instructions (typically 32 bits) • Regs: $0, $1, ..., $31 • %r0, %r1, ..., %r7 • Three-address architecture • Reg = 0: $0 • (n.a.) • RISC examples: MIPS, SPARC, • Return address: $31 • (n.a.) IBM/Motorola PowerPC, Compaq Alpha, ARM, SH4, HP-PA, ... CS 61C L29 Intel & Review (3) A Carle, Summer 2005 © UCB CS 61C L29 Intel & Review (4) A Carle, Summer 2005 © UCB MIPS vs. Intel 80x86 MIPS vs. Intel 80x86 • MIPS: “load-store architecture” • MIPS: “Three-address architecture” • Only Load/Store access memory; rest • Arithmetic-logic specify all 3 operands operations register-register; e.g., add $s0,$s1,$s2 # s0=s1+s2 lw $t0, 12($gp) add $s0,$s0,$t0 # s0=s0+Mem[12+gp] • Benefit: fewer instructions ⇒ performance • Benefit: simpler hardware ⇒ easier
    [Show full text]
  • Sebastian Hodge My Science Project 9/6/15 – 25/6/15
    Sebastian Hodge My Science Project 9/6/15 – 25/6/15 CPU Clock Speed and Performance Sebastian Hodge | CPU Clock Speed and Performance Contents Aim ................................................................................................................................................................ 3 Background Information ............................................................................................................................... 3 Introduction .............................................................................................................................................. 3 What is a CPU? What is clock speed? ....................................................................................................... 3 Do two processors with the same GHz have the same performance? ..................................................... 4 How can CPU clock speed be changed? .................................................................................................... 5 How is CPU performance measured? ....................................................................................................... 5 What is this experiment looking to find? .................................................................................................. 6 Hypothesis..................................................................................................................................................... 7 Materials/Equipment ...................................................................................................................................
    [Show full text]