An Empirical Analysis of Hardware Failures on a Million Consumer Pcs

Total Page:16

File Type:pdf, Size:1020Kb

An Empirical Analysis of Hardware Failures on a Million Consumer Pcs Cycles, Cells and Platters: An Empirical Analysis of Hardware Failures on a Million Consumer PCs Edmund B. Nightingale John R. Douceur Vince Orgovan Microsoft Research Microsoft Research Microsoft Corporation [email protected] [email protected] [email protected] Abstract two orders of magnitude after a first such crash occurs, and We present the first large-scale analysis of hardware failure DRAM failures are likely to recur in the same location. rates on a million consumer PCs. We find that many failures Three conditions motivate a hardware-failure study par- are neither transient nor independent. Instead, a large portion ticularly targeted at consumer machines. First, in contrast to of hardware induced failures are recurrent: a machine that servers, consumer machines tend to lack significant error- crashes from a fault in hardware is up to two orders of mag- resiliency features, such as ECC for memory and RAID for nitude more likely to crash a second time. For example, ma- disks, which increases the likelihood that a hardware er- chines with at least 30 days of accumulated CPU time over ror will have an impact on a machine’s operation. Second, an 8 month period had a 1 in 190 chance of crashing due whereas the failure of a single server in a data center can to a CPU subsystem fault. Further, machines that crashed often be masked by the application-level logic that parti- once had a probability of 1 in 3.3 of crashing a second tions tasks among machines, the failure of a single con- time. Our study examines failures due to faults within the sumer machine directly affects the user of that machine. CPU, DRAM and disk subsystems. Our analysis spans desk- Moreover, there are good reasons to expect that server-class tops and laptops, CPU vendor, overclocking, underclocking, and consumer-class machines will exhibit different failure generic vs. brand name, and characteristics such as machine characteristics: Consumer machines are built with cheaper speed and calendar age. Among our many results, we find components, and they run in harsher environments, with that CPU fault rates are correlated with the number of cycles wider temperature variation, greater mechanical stresses, executed, underclocked machines are significantly more reli- more ambient dust, and more frequent power cycling. Third, able than machines running at their rated speed, and laptops although there now exists a significant body of work on an- are more reliable than desktops. alyzing hardware failures in server machines [Bairavasun- daram et al. 2007, 2008; Kalyanakrishnam et al. 1999; Op- Categories and Subject Descriptors D.4.5 [Operating Sys- penheimer et al. 2003; Pinheiro et al. 2007; Schroeder and tems]: Reliability Gibson 2006, 2007; Schroeder et al. 2009; Xu et al. 1999], General Terms Measurement, reliability there is a dearth of information on consumer machines, even though they constitute the vast majority of machines sold Keywords Fault tolerance, hardware faults and used each year. Studying consumer machines brings new challenges not 1. Introduction present when studying server-class machines. Consumer We present the first large-scale analysis of hardware failure machines are geographically remote, widely distributed, and rates on consumer PCs by studying failures in the CPU, independently administered, which complicates data col- DRAM, and disk subsystems. We find that many failures are lection. There are no field reports from repair technicians neither transient (one-off) nor independent (memoryless). or outage logs filed by customers, as used by other stud- Instead, a large portion of hardware-induced failures are ies [Gray 1987, 1990; Kalyanakrishnam et al. 1999; Oppen- recurrent. Hardware crashes increase in likelihood by up to heimer et al. 2003; Schroeder and Gibson 2006; Xu et al. 1999]. The general lack of ECC precludes tracking ECC er- rors as a means for determining memory failures, as used in server studies [Constantinescu 2003; Schroeder et al. 2009]. Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed To address these challenges, we employ data sets from for profit or commercial advantage and that copies bear this notice and the full citation the Windows Error Reporting (WER) system [Glerum et al. on the first page. To copy otherwise, to republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. 2009], which was built to support diagnosis of software EuroSys’11, April 10–13, 2011, Salzburg, Austria. faults that occur in the field. Through post hoc analysis of re- Copyright c 2011 ACM 978-1-4503-0634-8/11/04. $10.00 ports from roughly one million machines, we are able to iso- 2. Prior work late several narrow classes of failures that are highly likely Ours is the first large-scale study of hardware failures in con- to have a root cause in hardware: machine-check exceptions sumer machines, and the first study of CPU subsystem fail- reported by the CPU, single-bit errors in the kernel region of ures on any class of machines. Furthermore, many of our DRAM, and OS-critical read failures in the disk subsystem. analyses have not been previously conducted, even for server This methodology has two unfortunate limitations. First, machines. The effect on failure rates of overclocking and un- since WER logs are generated only in response to a crash, derclocking, brand name vs. white box, and memory size our study is blind to hardware failures that do not lead to have not been previously studied. There is no prior work system crashes. In particular, the only DRAM bit errors that examining the relative frequency of intermittent vs. tran- cause system crashes are those that occur within the roughly sient hardware faults, nor in studying the spatial locality of 1.5% of memory that is occupied by kernel code pages. Er- DRAM failures. Some of our comparisons, such as desktop rors elsewhere in the memory may cause application crashes vs. laptop, are not even meaningful in server systems. or data corruption, but we cannot observe these. However, some of our findings mirror those in prior work. Second, our study sheds no light on the relative frequency Like Schroeder et al. [2009], who conducted detailed stud- of hardware-induced versus software-induced crashes. The ies of DRAM failure rates on server machines, we find that three types of failures we study constitute a minority of DRAM errors are far more likely to occur than would be ex- WER entries; however, most types of CPU, memory, and pected from studies of error rates induced by active radiation disk failures produce symptoms that are indistinguishable sources [Constantinescu 2002; Micron 1997; Ziegler et al. from kernel-software failures. On the other hand, such a re- 1998] or cosmic rays [Seifert et al. 2001; Ziegler et al. 1996]. sult would be relevant only for Windows machines, whereas Also, like Pinheiro et al. [2007] and Schroeder and Gibson absolute measures of hardware failures are relevant for any [2007], who examined disk failure rates in large data-center operating system, or at least any OS that relies on a function- installations, we find that disk MTTF times are much lower ing CPU, error-free memory, and a responsive disk. than those specified on disk data sheets. Our findings on Despite these limitations, our study has found a number conditional failure probability mirror the findings of recent of interesting results. For instance, even small degrees of studies of DRAM failures [Schroeder et al. 2009] and disk- overclocking significantly degrade machine reliability, and subsystem failures [Jiang et al. 2008] on server machines: an small degrees of underclocking improve reliability over run- increase of up to two orders of magnitude in observed failure ning at rated speed. We also find that faster CPUs tend to rates after a first failure occurs. become faulty more quickly than slower CPUs, and laptops Some of our findings contrast with those of prior studies. have lower failure rates than desktops. Beyond our results, In studying the correlation between age and DRAM failure this study serves to inform the community that hardware rates, Schroeder et al. [2009] found no infant mortality and faults on consumer machines are not rare, independent, or no increasing error rates at very old ages; by contrast, we find always transient. While prior work focused on measuring minor evidence for both of these phenomena, consistent with hardware faults that may or may not cause an OS failure, a bathtub curve. Constantinescu [2003], in a study of ECC our study is the first to measure observed OS failure rates errors on 193 servers, concluded that the risk of intermittent due to faulty hardware. faults increases as chip densities and frequencies increase; In the sequel, we analyze the probability of machine fail- by contrast, we find that memory size (which roughly corre- ure from each failure type, as well as the conditional prob- sponds to density) has only weak correlation to failure rate, ability that a failure will recur (§4). We analyze the spatial and CPU frequency has no effect on the rate of failures per locality of DRAM failures, which informs conjectures about CPU cycle for non-overclocked machines. the underlying fault. We analyze failure behavior of various Much of the prior work in this area addresses issues that machine classes, by partitioning machines into populations our study does not. Our data is unable to conclusively ad- of overclocked vs. non-overclocked, underclocked vs. rated- dress the relative frequency of hardware versus software fail- speed, white box (generic) vs. brand name, and desktops vs. ures, but the coarse-grained division between hardware, soft- laptops (§5).
Recommended publications
  • Computer Service Technician- CST Competency Requirements
    Computer Service Technician- CST Competency Requirements This Competency listing serves to identify the major knowledge, skills, and training areas which the Computer Service Technician needs in order to perform the job of servicing the hardware and the systems software for personal computers (PCs). The present CST COMPETENCIES only address operating systems for Windows current version, plus three older. Included also are general common Linux and Apple competency information, as proprietary service contracts still keep most details specific to in-house service. The Competency is written so that it can be used as a course syllabus, or the study directed towards the education of individuals, who are expected to have basic computer hardware electronics knowledge and skills. Computer Service Technicians must be knowledgeable in the following technical areas: 1.0 SAFETY PROCEDURES / HANDLING / ENVIRONMENTAL AWARENESS 1.1 Explain the need for physical safety: 1.1.1 Lifting hardware 1.1.2 Electrical shock hazard 1.1.3 Fire hazard 1.1.4 Chemical hazard 1.2 Explain the purpose for Material Safety Data Sheets (MSDS) 1.3 Summarize work area safety and efficiency 1.4 Define first aid procedures 1.5 Describe potential hazards in both in-shop and in-home environments 1.6 Describe proper recycling and disposal procedures 2.0 COMPUTER ASSEMBLY AND DISASSEMBLY 2.1 List the tools required for removal and installation of all computer system components 2.2 Describe the proper removal and installation of a CPU 2.2.1 Describe proper use of Electrostatic Discharge
    [Show full text]
  • Comptia A+ Complete Study Guide A+ Essentials (220-601) Exam Objectives
    4830bperf.fm Page 1 Thursday, March 8, 2007 10:03 AM CompTIA A+ Complete Study Guide A+ Essentials (220-601) Exam Objectives OBJECTIVE CHAPTER Domain 1.0 Personal Computer Components 1.1 Identify the fundamental principles of using personal computers 1 1.2 Install, configure, optimize and upgrade personal computer components 2 1.3 Identify tools, diagnostic procedures and troubleshooting techniques for personal computer components 2 1.4 Perform preventative maintenance on personal computer components 2 Domain 2.0 Laptops and Portable Devices 2.1 Identify the fundamental principles of using laptops and portable devices 3 2.2 Install, configure, optimize and upgrade laptops and portable devices 3 2.3 Identify tools, basic diagnostic procedures and troubleshooting techniques for laptops and portable devices 3 2.4 Perform preventative maintenance on laptops and portable devices 3 Domain 3.0 Operating Systems 3.1 Identify the fundamentals of using operating systems 4 3.2 Install, configure, optimize and upgrade operating systems 5 3.3 Identify tools, diagnostic procedures and troubleshooting techniques for operating systems 6 3.4 Perform preventative maintenance on operating systems 6 Domain 4.0 Printers and Scanners 4.1 Identify the fundamental principles of using printers and scanners 7 4.2 Identify basic concepts of installing, configuring, optimizing and upgrading printers and scanners 7 4.3 Identify tools, basic diagnostic procedures and troubleshooting techniques for printers and scanners 7 Domain 5.0 Networks 5.1 Identify the fundamental
    [Show full text]
  • Overclocking Assistant for the Intel® Desktop Board DZ87KLT-75K
    Overclocking Assistant for the Intel® Desktop Board DZ87KLT-75K June 2013 Overclocking Assistant for the Intel® Desktop Board DZ87KLT-75K WARNING Altering clock frequency and/or voltage may (i) reduce system stability and useful life of the system and processor; (ii) cause the processor and other system components to fail; (iii) cause reductions in system performance; (iv) cause additional heat or other damage; and (v) affect system data integrity. Intel has not tested and does not warranty the operation of the processor beyond its specifications. WARNING Altering PC memory frequency and/or voltage may (i) reduce system stability and useful life of the system, memory and processor; (ii) cause the processor and other system components to fail; (iii) cause reductions in system performance; (iv) cause additional heat or other damage; and (v) affect system data integrity. Intel assumes no responsibility that the memory included, if used with altered clock frequencies and/or voltages, will be fit for any particular purpose. Check with the memory manufacturer for warranty and additional details. INFORMATION IN THIS DOCUMENT IS PROVIDED IN CONNECTION WITH INTEL® PRODUCTS. NO LICENSE, EXPRESS OR IMPLIED, BY ESTOPPEL OR OTHERWISE, TO ANY INTELLECTUAL PROPERTY RIGHTS IS GRANTED BY THIS DOCUMENT. EXCEPT AS PROVIDED IN INTEL’S TERMS AND CONDITIONS OF SALE FOR SUCH PRODUCTS, INTEL ASSUMES NO LIABILITY WHATSOEVER, AND INTEL DISCLAIMS ANY EXPRESS OR IMPLIED WARRANTY, RELATING TO SALE AND/OR USE OF INTEL PRODUCTS INCLUDING LIABILITY OR WARRANTIES RELATING TO FITNESS FOR A PARTICULAR PURPOSE, MERCHANTABILITY, OR INFRINGEMENT OF ANY PATENT, COPYRIGHT OR OTHER INTELLECTUAL PROPERTY RIGHT. Intel products are not intended for use in medical, life saving, or life sustaining applications.
    [Show full text]
  • Temperature Adaptive Computing in a Real PC
    TEAPC: Temperature Adaptive Computing in a Real PC Augustus K. Uht and Richard J. Vaccaro University of Rhode Island Microarchitecture Research Institute Department of Electrical and Computer Engineering 4 East Alumni Ave. Kingston, RI 02864, USA {[email protected], [email protected]} Abstract* Herein we present an adaptive prototype, TEAPC, based on the standard IBM/Intel PC architecture and TEAPC is an IBM/Intel-standard PC realization realizing underclocking, that is, operating the system of a CPU temperature-adaptive feedback-control greatly below its specified frequency. We have system. The control system adjusts the CPU frequency implemented the entire TEAPC control system in and/or voltage to maintain a constant set-point software, in a Windows application. It runs in a temperature. TEAPC dynamically adapts to changing normal application mode on Windows 2000, CPU computation loads, as well as any other system multitasking normally with other applications. CPU phenomenon that affects the CPU temperature. For frequency and core voltage are changed dynamically, example, with the specific TEAPC hardware, should based on the output of a feedback control system using the CPU cooling fan fail, the control system detects the CPU chip temperature as its input. The control the increase in CPU temperature and automatically loop and underclocking models should be usable for reduces the CPU frequency and voltage to keep the any PC to be built, as well as many that already exist. CPU from overheating. In such a circumstance the This paper’s contributions are as follows: system continues to operate. All of the adaptation is Demonstration of temperature control of a COTS done dynamically, at runtime, on an unmodified (Commercial Off-The-Shelf) based PC via a modern standard operating system (Windows 2000) with a feedback control system; On-demand reliability and purely software-implemented feedback-control system.
    [Show full text]
  • Comptia A+ Certification Exam Prep
    Course 445 CompTIA A+ Certification Exam Prep 445/CN/H.3/709/H.2 Acknowledgments The author would like to thank the following people for their valuable help with this course: Teresa Shinn Carl Waldron Tim Watts Esteban Delgado Dave O’Neal Shirley Auguste Alicia Richardson © Learning Tree International, Inc. All rights reserved. Not to be reproduced without prior written consent. © LEARNING TREE INTERNATIONAL, INC. All rights reserved. All trademarked product and company names are the property of their respective trademark holders. No part of this publication may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, electronic, mechanical, photocopying, recording or otherwise, or translated into any language, without the prior written permission of the publisher. Copying software used in this course is prohibited without the express permission of Learning Tree International, Inc. Making unauthorized copies of such software violates federal copyright law, which includes both civil and criminal penalties. Introduction and Overview Course Objectives The main objective of this course is to Prepare you to pass the CompTIA A+ 220-901 and 220-902 exams To accomplish this goal, you will learn how to Safely disassemble and reassemble a complete PC system Install and configure common PC motherboards and adapter cards Use an organized strategy to identify and troubleshoot common PC problems Select and install the correct type of RAM memory needed to upgrade a PC system Install and configure data storage systems, including hard disks, solid state drives (SSD), CD, and DVD-ROM drives Identify fundamental components of laptops and mobile devices and describe how to troubleshoot common problems © Learning Tree International, Inc.
    [Show full text]
  • Error Handling and Energy Estimation Framework for Error Resilient Near-Threshold Computing Rengarajan Ragavan
    Error Handling and Energy Estimation Framework For Error Resilient Near-Threshold Computing Rengarajan Ragavan To cite this version: Rengarajan Ragavan. Error Handling and Energy Estimation Framework For Error Resilient Near- Threshold Computing. Embedded Systems. Rennes 1, 2017. English. tel-01636803 HAL Id: tel-01636803 https://hal.inria.fr/tel-01636803 Submitted on 17 Nov 2017 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. No d’ordre : ANNÉE 2017 THÈSE / UNIVERSITÉ DE RENNES 1 sous le sceau de l’Université Bretagne Loire pour le grade de DOCTEUR DE L’UNIVERSITÉ DE RENNES 1 Mention : Traitement du Signal et Télécommunications École doctorale MATISSE présentée par Rengarajan RAGAVAN préparée à l’unité de recherche UMR6074 IRISA Institut de recherche en informatique et systèmes aléatoires - CAIRN École Nationale Supérieure des Sciences Appliquées et de Technologie Thèse soutenue à Lannion le 22 septembre 2017 devant le jury composé de : Edith BEIGNÉ Error Handling and Directrice de Recherche à CEA Leti / rapporteur Energy Estimation Patrick GIRARD Directrice de Recherche à CNRS, LIRMM / rappor- teur Framework For Lorena ANGHEL Error Resilient Professeur à Grenoble INP, TIMA / examinateur Daniel MÉNARD Near-Threshold Professeur à INSA Rennes, IETR / examinateur Olivier SENTIEYS Computing Professeur à l’Université de Rennes 1 / directeur de thèse Cédric KILLIAN Maître de Conférences à l’Université de Rennes 1 / co-directeur de thèse To My Parents, My Brother, and My Teachers iii Acknowledgements I sincerely express my gratitude to my thesis supervisor Prof.
    [Show full text]
  • Dynamic Voltage and Frequency Scaling with Multi-Clock Distribution Systems on SPARC Core" (2009)
    Rochester Institute of Technology RIT Scholar Works Theses Thesis/Dissertation Collections 5-1-2009 Dynamic voltage and frequency scaling with multi- clock distribution systems on SPARC core Michael Nasri Michael Follow this and additional works at: http://scholarworks.rit.edu/theses Recommended Citation Michael, Michael Nasri, "Dynamic voltage and frequency scaling with multi-clock distribution systems on SPARC core" (2009). Thesis. Rochester Institute of Technology. Accessed from This Thesis is brought to you for free and open access by the Thesis/Dissertation Collections at RIT Scholar Works. It has been accepted for inclusion in Theses by an authorized administrator of RIT Scholar Works. For more information, please contact [email protected]. Dynamic Voltage and Frequency Scaling with Multi-Clock Distribution Systems on SPARC Core by Michael Nasri Michael A Thesis Submitted in Partial Fulfillment of the Requirements for the Degree of Master of Science in Computer Engineering Supervised by Dr. Dhireesha Kudithipudi Department of Computer Engineering Kate Gleason College of Engineering Rochester Institute of Technology Rochester, NY May 2009 Approved By: Dr. Dhireesha Kudithipudi Assistant Professor, RIT Department of Computer Engineering Primary Advisor Dr. Ken Hsu Professor, RIT Department of Computer Engineering Dr. Muhammad Shaaban Associate Professor, RIT Department of Computer Engineering Thesis Release Permission Form Rochester Institute of Technology Kate Gleason College of Engineering Title: Dynamic Voltage and Frequency Scaling with Multi-Clock Distribution Systems on SPARC Core I, Michael Nasri Michael, hereby grant permission to the Wallace Memorial Library to reproduce my thesis in whole or part. Michael Nasri Michael Date Dedication To my parents for their pride and encouragement To my brother and sisters for their love and support Acknowledgements Thanks to Dr.
    [Show full text]
  • Geosciences and Climate on Marvell Thunderx2 White Paper
    WHITE PAPER Geosciences and Climate on Marvell ThunderX2 Server Processor Business Unit Marvell June 2019 ABSTRACT Earthquakes and climate change affect everyone on our planet. Increases in computational capability are letting scientists simulate these phenomena with increasingly finer resolution grids and with greater accuracy and urgency. In this white paper, we present why the Marvell® ThunderX2® Arm®-based server processor provides significant value for end users deploying applications in the field of weather modeling and geosciences. Marvell ThunderX2 – Designed for High-Performance Computing Before we dig deeper in to the ThunderX2 architecture, it is important to understand a typical way of categorizing high-performance computing (HPC) deployments. High-performance computing is broadly divided into two categoriesi. Capability computing refers to using a large- scale HPC installation to solve a single problem in the shortest possible time, for example simulating weather models on a Tier-0 HPC system. Capacity computing refers to optimizing system efficiency to solve as many mid-size or smaller problems as possible at the same time at the lowest possible cost, for example automobile manufacturers using rented (on-demand) HPC resources to simulate numerous drive models for their products. It is clear from the definitions above that an HPC customer must incorporate components of cost/CPU, # of CPUs and performance (execution time)/CPU. Thus, the cost of an HPC application can be summarized as cost α CPU hours where CPU hours = #cores x execution time The other important factor driving design decisions of an HPC system is the application suite running on a cluster. Although capability computing targets application runs with the lowest execution time, excessive application scaling may deliver diminishing returns in performance improvement while linearly increasing CPU-hours.
    [Show full text]
  • Error Handling and Energy Estimation for Error Resilient Near-Threshold Computing Rengarajan Ragavan
    Error handling and energy estimation for error resilient near-threshold computing Rengarajan Ragavan To cite this version: Rengarajan Ragavan. Error handling and energy estimation for error resilient near-threshold com- puting. Hardware Architecture [cs.AR]. Université Rennes 1, 2017. English. NNT : 2017REN1S038. tel-01654476 HAL Id: tel-01654476 https://tel.archives-ouvertes.fr/tel-01654476 Submitted on 4 Dec 2017 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. No d’ordre : ANNÉE 2017 THÈSE / UNIVERSITÉ DE RENNES 1 sous le sceau de l’Université Bretagne Loire pour le grade de DOCTEUR DE L’UNIVERSITÉ DE RENNES 1 Mention : Traitement du Signal et Télécommunications École doctorale MATISSE présentée par Rengarajan RAGAVAN préparée à l’unité de recherche UMR6074 IRISA Institut de recherche en informatique et systèmes aléatoires - CAIRN École Nationale Supérieure des Sciences Appliquées et de Technologie Thèse soutenue à Lannion le 22 septembre 2017 devant le jury composé de : Edith BEIGNÉ Error Handling and Directrice de Recherche à CEA Leti / rapporteur
    [Show full text]
  • Power Management in Tizen and Linux Kernel (Tizen 2.3)
    Power Management in Tizen and Linux Kernel (Tizen 2.3) Minsoo Ryu Real-Time Computing and Communications Lab. Hanyang University [email protected] Real-Time Computing and Communications Lab., Hanyang University http://rtcc.hanyang.ac.kr Outline Introduction ACPI (Advanced Configuration Power Interface) DVFS (Dynamic Voltage Frequency Scaling) Power Management in Tizen Lab Assignment Real-Time Computing and Communications Lab., Hanyang University http://rtcc.hanyang.ac.kr 22 Why Power Management? Battery-operated devices . Smartphones, digital cameras, and laptops use batteries . Power savings and battery run times are paramount Increasing energy use of servers . Servers and data centers in U.S. reached 1.5% of the national electricity consumption in 2006 . The electricity consumption of servers doubled between 2000 and 2006 Real-Time Computing and Communications Lab., Hanyang University http://rtcc.hanyang.ac.kr 33 Power Consumption in Smartphones Real-Time Computing and Communications Lab., Hanyang University http://rtcc.hanyang.ac.kr 44 Power Consumption of Memory In a server system . Memory consumes 19% of system power on average . Some work notes up to 40% of total system power Real-Time Computing and Communications Lab., Hanyang University http://rtcc.hanyang.ac.kr 55 Idle Power Consumption Only 30% of servers in data centers are fully utilized while keeping the other 70% in idle state . Idle servers consume between 60% and 66% of the peak load power consumption Real-Time Computing and Communications Lab., Hanyang University http://rtcc.hanyang.ac.kr 66 Two Dimensions of Power Management Power management when the system is idle . Select the most efficient idle state Power management when the system is active .
    [Show full text]
  • Beyond CPU: Considering Memory Power Consumption of Software
    Beyond CPU: Considering Memory Power Consumption of Software Hayri Acar1, Gülfem I. Alptekin2, Jean-Patrick Gelas3 and Parisa Ghodous1 1LIRIS, University of Lyon, Lyon, France 2Galatasaray University, Istanbul, Turkey 3ENS Lyon, LIP, UMR 5668, Lyon, France Keywords: Power Consumption, Sustainable Software, Energy Efficiency, Green IT. Abstract: ICTs (Information and Communication Technologies) are responsible around 2% of worldwide greenhouse gas emissions (Gartner, 2007). And according to the Intergovernmental Panel on Climate Change (IPPC) recent reports, CO2 emissions due to ICTs are increasing widely. For this reason, many works tried to propose various tools to estimate the energy consumption due to software in order to reduce carbon footprint. However, these studies, in the majority of cases, takes into account only the CPU and neglects all others components. Whereas, the trend towards high-density packaging and raised memory involve a great increased of power consumption caused by memory and maybe memory can become the largest power consumer in servers. In this paper, we model and then estimate the power consumed by CPU and memory due to the execution of a software. Thus, we perform several experiments in order to observe the behavior of each component. 1 INTRODUCTION more possible to obtain accurate and efficient results for energy consumption. However, using these types ICTs (Information and Communication of devices is complicated because it is necessary to Technologies) are responsible around 2% of have these devices and connect them to different worldwide greenhouse gas emissions (Gartner, components. What is more with this method, it is 2007). And according to the Intergovernmental Panel impossible to measure the energy consumed by on Climate Change (IPPC) recent reports, CO2 virtual machines and applications on process.
    [Show full text]
  • TEAPC: Adaptive Computing and Underclocking in a Real PC Augustus K
    From: The Proceedings of the First IBM P=ac2 Conference, October 6-8, 2004,IBM T.J. Watson Research Center, Yorktown Heights, NY, USA. Copyright 2004, A.K. Uht and R.J. Vaccaro. TEAPC: Adaptive Computing and Underclocking in a Real PC Augustus K. Uht and Richard J. Vaccaro University of Rhode Island Microarchitecture Research Institute Department of Electrical and Computer Engineering 4 East Alumni Ave. Kingston, RI 02864, USA {[email protected], [email protected]} real-time internal temperature. Since this is the major Abstract component of variation of timing within an Intel TEAPC is an IBM/Intel-standard PC realization of microprocessor (the CPU’s core voltage is highly the TEAtime performance “maximizing” adaptive regulated and approximately constant), it mirrors the computing algorithm, giving performance beyond original TEAtime philosophy. Further, it is possible worst-case-specifications. TEAPC goes beyond the to base the real-time performance on the real-time TEAtime algorithm by adapting to the current CPU power consumption of the entire PC; the latter is load. It is also the first machine to use extensive partially addressed in this paper. underclocking for disaster tolerance, low power consumption and high reliability. This is all done We have implemented the entire control system in dynamically, at runtime, on an unmodified standard software, in a Windows application. It runs in a operating system (Windows 2000) with a purely normal application mode on Windows 2000, software-implemented feedback-control based multitasking normally with other applications. CPU algorithm. frequency and core voltage are changed dynamically, based on the output of a feedback control system 1 Introduction using the CPU chip temperature as its primary input.
    [Show full text]