The TURBO Diaries: Application-Controlled Frequency
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Improving Resource Utilization in Heterogeneous CPU-GPU Systems
Improving Resource Utilization in Heterogeneous CPU-GPU Systems A Dissertation Presented to the Faculty of the School of Engineering and Applied Science University of Virginia In Partial Fulfillment of the requirements for the Degree Doctor of Philosophy (Computer Engineering) by Michael Boyer May 2013 c 2013 Michael Boyer Abstract Graphics processing units (GPUs) have attracted enormous interest over the past decade due to substantial increases in both performance and programmability. Programmers can potentially leverage GPUs for substantial performance gains, but at the cost of significant software engineering effort. In practice, most GPU applications do not effectively utilize all of the available resources in a system: they either fail to use use a resource at all or use a resource to less than its full potential. This underutilization can hurt both performance and energy efficiency. In this dissertation, we address the underutilization of resources in heterogeneous CPU-GPU systems in three different contexts. First, we address the underutilization of a single GPU by reducing CPU-GPU interaction to improve performance. We use as a case study a computationally-intensive video-tracking application from systems biology. Because of the high cost of CPU-GPU coordination, our initial, straightforward attempts to accelerate this application failed to effectively utilize the GPU. By leveraging some non-obvious optimization strategies, we significantly decreased the amount of CPU-GPU interaction and improved the performance of the GPU implementation by 26x relative to the best CPU implementation. Based on the lessons we learned, we present general guidelines for optimizing GPU applications as well as recommendations for system-level changes that would simplify the development of high-performance GPU applications. -
Analysis of Body Bias Control Using Overhead Conditions for Real Time Systems: a Practical Approach∗
IEICE TRANS. INF. & SYST., VOL.E101–D, NO.4 APRIL 2018 1116 PAPER Analysis of Body Bias Control Using Overhead Conditions for Real Time Systems: A Practical Approach∗ Carlos Cesar CORTES TORRES†a), Nonmember, Hayate OKUHARA†, Student Member, Nobuyuki YAMASAKI†, Member, and Hideharu AMANO†, Fellow SUMMARY In the past decade, real-time systems (RTSs), which must in RTSs. These techniques can improve energy efficiency; maintain time constraints to avoid catastrophic consequences, have been however, they often require a large amount of power since widely introduced into various embedded systems and Internet of Things they must control the supply voltages of the systems. (IoTs). The RTSs are required to be energy efficient as they are used in embedded devices in which battery life is important. In this study, we in- Body bias (BB) control is another solution that can im- vestigated the RTS energy efficiency by analyzing the ability of body bias prove RTS energy efficiency as it can manage the tradeoff (BB) in providing a satisfying tradeoff between performance and energy. between power leakage and performance without affecting We propose a practical and realistic model that includes the BB energy and the power supply [4], [5].Itseffect is further endorsed when timing overhead in addition to idle region analysis. This study was con- ducted using accurate parameters extracted from a real chip using silicon systems are enabled with silicon on thin box (SOTB) tech- on thin box (SOTB) technology. By using the BB control based on the nology [6], which is a novel and advanced fully depleted sili- proposed model, about 34% energy reduction was achieved. -
Application-Controlled Frequency Scaling Explained
The TURBO Diaries: Application-controlled Frequency Scaling Explained Jons-Tobias Wamhoff, Stephan Diestelhorst, and Christof Fetzer, Technische Universät Dresden; Patrick Marlier and Pascal Felber, Université de Neuchâtel; Dave Dice, Oracle Labs https://www.usenix.org/conference/atc14/technical-sessions/presentation/wamhoff This paper is included in the Proceedings of USENIX ATC ’14: 2014 USENIX Annual Technical Conference. June 19–20, 2014 • Philadelphia, PA 978-1-931971-10-2 Open access to the Proceedings of USENIX ATC ’14: 2014 USENIX Annual Technical Conference is sponsored by USENIX. The TURBO Diaries: Application-controlled Frequency Scaling Explained Jons-Tobias Wamhoff Patrick Marlier Dave Dice Stephan Diestelhorst Pascal Felber Christof Fetzer Technische Universtat¨ Dresden, Germany Universite´ de Neuchatel,ˆ Switzerland Oracle Labs, USA Abstract these features from an application as needed. Examples in- Most multi-core architectures nowadays support dynamic volt- clude: accelerating the execution of key sections of code on age and frequency scaling (DVFS) to adapt their speed to the the critical path of multi-threaded applications [9]; boosting system’s load and save energy. Some recent architectures addi- time-critical operations or high-priority threads; or reducing tionally allow cores to operate at boosted speeds exceeding the the energy consumption of applications executing low-priority nominal base frequency but within their thermal design power. threads. Furthermore, workloads specifically designed to run In this paper, we propose a general-purpose library that on processors with heterogeneous cores (e.g., few fast and allows selective control of DVFS from user space to accelerate many slow cores) may take additional advantage of application- multi-threaded applications and expose the potential of hetero- level frequency scaling. -
Dynamic Voltage/Frequency Scaling and Power-Gating of Network-On-Chip with Machine Learning
Dynamic Voltage/Frequency Scaling and Power-Gating of Network-on-Chip with Machine Learning A thesis presented to the faculty of the Russ College of Engineering and Technology of Ohio University In partial fulfillment of the requirements for the degree Master of Science Mark A. Clark May 2019 © 2019 Mark A. Clark. All Rights Reserved. 2 This thesis titled Dynamic Voltage/Frequency Scaling and Power-Gating of Network-on-Chip with Machine Learning by MARK A. CLARK has been approved for the School of Electrical Engineering and Computer Science and the Russ College of Engineering and Technology by Avinash Karanth Professor of Electrical Engineering and Computer Science Dennis Irwin Dean, Russ College of Engineering and Technology 3 Abstract CLARK, MARK A., M.S., May 2019, Electrical Engineering Dynamic Voltage/Frequency Scaling and Power-Gating of Network-on-Chip with Machine Learning (89 pp.) Director of Thesis: Avinash Karanth Network-on-chip (NoC) continues to be the preferred communication fabric in multicore and manycore architectures as the NoC seamlessly blends the resource efficiency of the bus with the parallelization of the crossbar. However, without adaptable power management the NoC suffers from excessive static power consumption at higher core counts. Static power consumption will increase proportionally as the size of the NoC increases to accommodate higher core counts in the future. NoC also suffers from excessive dynamic energy as traffic loads fluctuate throughout the execution of an application. Power- gating (PG) and Dynamic Voltage and Frequency Scaling (DVFS) are two highly effective techniques proposed in literature to reduce static power and dynamic energy in the NoC respectively. -
Real-Time Dynamic Voltage Scaling for Low-Power Embedded Operating Systems£
Real-Time Dynamic Voltage Scaling for Low-Power Embedded Operating Systems£ Padmanabhan Pillai and Kang G. Shin Real-Time Computing Laboratory Department of Electrical Engineering and Computer Science The University of Michigan Ann Arbor, MI 48109-2122, U.S.A. pillai,kgshin @eecs.umich.edu ABSTRACT ful microprocessors running sophisticated, intelligent control soft- In recent years, there has been a rapid and wide spread of non- ware in a vast array of devices including digital camcorders, cellu- traditional computing platforms, especially mobile and portable com- lar phones, and portable medical devices. puting devices. As applications become increasingly sophisticated and processing power increases, the most serious limitation on these Unfortunately, there is an inherent conflict in the design goals be- devices is the available battery life. Dynamic Voltage Scaling (DVS) hind these devices: as mobile systems, they should be designed to has been a key technique in exploiting the hardware characteristics maximize battery life, but as intelligent devices, they need powerful of processors to reduce energy dissipation by lowering the supply processors, which consume more energy than those in simpler de- voltage and operating frequency. The DVS algorithms are shown to vices, thus reducing battery life. In spite of continuous advances in be able to make dramatic energy savings while providing the nec- semiconductor and battery technologies that allow microprocessors essary peak computation power in general-purpose systems. How- to provide much greater computation per unit of energy and longer ever, for a large class of applications in embedded real-time sys- total battery life, the fundamental tradeoff between performance tems like cellular phones and camcorders, the variable operating and battery life remains critically important. -
Learning-Directed Dynamic Voltage and Frequency Scaling Scheme with Adjustable Performance for Single-Core and Multi-Core Embedded and Mobile Systems †
sensors Article Learning-Directed Dynamic Voltage and Frequency Scaling Scheme with Adjustable Performance for Single-Core and Multi-Core Embedded and Mobile Systems † Yen-Lin Chen 1,* , Ming-Feng Chang 2, Chao-Wei Yu 1 , Xiu-Zhi Chen 1 and Wen-Yew Liang 1 1 Department of Computer Science and Information Engineering, National Taipei University of Technology, Taipei 10608, Taiwan; [email protected] (C.-W.Y.); [email protected] (X.-Z.C.); [email protected] (W.-Y.L.) 2 MediaTek Inc., Hsinchu 30078, Taiwan; [email protected] * Correspondence: [email protected]; Tel.: +886-2-27712171 (ext. 4239) † This paper is an expanded version of “Learning-Directed Dynamic Volt-age and Frequency Scaling for Computation Time Prediction” published in Proceedings of 2011 IEEE 10th International Conference on Trust, Security and Privacy in Computing and Communications, Changsha, China, 16–18 November 2011. Received: 6 August 2018; Accepted: 8 September 2018; Published: 12 September 2018 Abstract: Dynamic voltage and frequency scaling (DVFS) is a well-known method for saving energy consumption. Several DVFS studies have applied learning-based methods to implement the DVFS prediction model instead of complicated mathematical models. This paper proposes a lightweight learning-directed DVFS method that involves using counter propagation networks to sense and classify the task behavior and predict the best voltage/frequency setting for the system. An intelligent adjustment mechanism for performance is also provided to users under various performance requirements. The comparative experimental results of the proposed algorithms and other competitive techniques are evaluated on the NVIDIA JETSON Tegra K1 multicore platform and Intel PXA270 embedded platforms. -
Energy Proportional Computing in Commercial Fpgas with Adaptive
Energy proportional computing in Commercial FPGAs with Adaptive Voltage Scaling Jose Nunez-Yanez Department of Electrical and Electronic Engineering University of Bristol, UK +44 117 3315128 [email protected] ABSTRACT an open-loop configuration. However, worst case variability is Voltage and frequency adaptation can be used to create energy rarely the case. For this reason, this paper investigates Adaptive proportional systems in which energy usage adapts to the amount Voltage Scaling (AVS) in which run-time monitoring of of work to be done in the available time. Closed-loop voltage and performance variability in the silicon is used together with system frequency scaling can also take into account process and characterization to influence the voltage and the frequency on the temperature variations in addition to system load and this removes fly in a closed-loop configuration. a significant proportion of the margins used by device The rest of the paper is structured as follows. Section 2 describes manufacturers. This paper explores the capabilities of commercial related work. Section 3 presents the hardware platform used in FPGAs to use closed-loop adaptive voltage scaling to improve this research while section 4 introduces the design flow that their energy and performance profiles beyond nominal. An embeds the AVS capabilities in the user design. Section 5 adaptive power architecture based on a modified design flow is presents the power adaptive architecture based on the novel in- created with in-situ detectors and dynamic reconfiguration of situ detectors. Section 6 presents and discusses the results clock management resources. The results of deploying AVS in focusing on power and energy measurements. -
AMD's Early Processor Lines, up to the Hammer Family (Families K8
AMD’s early processor lines, up to the Hammer Family (Families K8 - K10.5h) Dezső Sima October 2018 (Ver. 1.1) Sima Dezső, 2018 AMD’s early processor lines, up to the Hammer Family (Families K8 - K10.5h) • 1. Introduction to AMD’s processor families • 2. AMD’s 32-bit x86 families • 3. Migration of 32-bit ISAs and microarchitectures to 64-bit • 4. Overview of AMD’s K8 – K10.5 (Hammer-based) families • 5. The K8 (Hammer) family • 6. The K10 Barcelona family • 7. The K10.5 Shanghai family • 8. The K10.5 Istambul family • 9. The K10.5-based Magny-Course/Lisbon family • 10. References 1. Introduction to AMD’s processor families 1. Introduction to AMD’s processor families (1) 1. Introduction to AMD’s processor families AMD’s early x86 processor history [1] AMD’s own processors Second sourced processors 1. Introduction to AMD’s processor families (2) Evolution of AMD’s early processors [2] 1. Introduction to AMD’s processor families (3) Historical remarks 1) Beyond x86 processors AMD also designed and marketed two embedded processor families; • the 2900 family of bipolar, 4-bit slice microprocessors (1975-?) used in a number of processors, such as particular DEC 11 family models, and • the 29000 family (29K family) of CMOS, 32-bit embedded microcontrollers (1987-95). In late 1995 AMD cancelled their 29K family development and transferred the related design team to the firm’s K5 effort, in order to focus on x86 processors [3]. 2) Initially, AMD designed the Am386/486 processors that were clones of Intel’s processors. -
The Dynamic Voltage and Frequency Scaling Based on the On-Chip Microcontroller System
Journal of Theoretical and Applied Information Technology 10th May 2013. Vol. 51 No.1 © 2005 - 2013 JATIT & LLS. All rights reserved. ISSN: 1992-8645 www.jatit.org E-ISSN: 1817-3195 THE DYNAMIC VOLTAGE AND FREQUENCY SCALING BASED ON THE ON-CHIP MICROCONTROLLER SYSTEM 1,2TIEFENG LI, 1CAIWEN MA, 1WENHUA LI 1 Xi'an Institute of Optics and Precision Mechanics of Chinese Academy of Sciences, Xi’an, 710119,China 2The Graduate University of Chinese Academy of Sciences ,Beijing, 100049, China E-mail: [email protected] , [email protected] , [email protected] ABSTRACT With the rapid increase of complexity and size of the on-chip microcontroller system (OCMS), the power consumption issue for the OMCS is increasingly becoming critical and needs to be solved quickly. Because the CPU is often the major power consumer in the OCMS, so an important strategy to achieve energy saving is via the dynamic voltage and frequency scaling (DVFS), which can enable a processor to operate at a range of voltages and frequencies. However, it needs to be emphasized that the conventional DVFS is fully executed by the software scheduler in the operating system, and its main drawback is that the scheduler can’t accurately track the performance requirements of CPU when the dormant frequency of CPU is increasing continuously. In this paper, we firstly present a typical hardware DVFS architecture, which can automatically carry out DVFS without the software scheduler involvement. Therefore, it avoids increasing the software's workload and reduces the power consumption. Keywords: Power Consumption, DVFS, Software Scheduler, Performance Requirement 1. -
LG8: (E) Engineering and Physical Considerations
LG8: (E) Engineering and Physical Considerations Topics: Power consumption, scaling, size, logical effort and performance limits. LG8.1 - E - 90 Nanometer Gate Length. LG8.2 - E - Power Consumption LG8.3 - E - Dynamic Power Gating LG8.4 - E - Dynamic Frequency Scaling LG8.5 - E - Dynamic Voltage Scaling LG8.6 - E - Logical Effort LG8.7 - E - Information Flux What is the Silicon End Point ? 1 LG8.1 - E - 90 Nanometer Gate Length. The mainstream VLSI technology in the period 2004-2008. Parameters from a 90 nanometer standard cell library: Parameter Value Unit Drawn Gate Length 0.08 µm Metal Layers 6 to 9 layers Max Gate Density 400K gates/mm2 Track Width 0.25 µm Track Spacing 0.25 µm Tracking Capacitance 1 fF/mm Core Supply Voltage 0.9 to 1.4 V FO4 Delay 51 ps Typical processor core: 200k gates + 4 RAMs: one square millimeter. Typical SoC chip area is 50-100 mm2 20-40 million gates. Actual gate and transistor count higher owing to custom blocks (RAMs mainly). Now the industry is moving to 45 nanometer. 2007: Dual-core Intel Itanium2: 2 billion transistors. http://en.wikipedia.org/wiki/Moore’s_law 2 LG8.2- E - Power Consumption P = V × I = E × f I (current) = Static Current + Dynamic Current. Early CMOS (VCC= 5 volts): negligible static current. Today it’s 30 % of dynamic. Dynamic current = Short circuit current + Charge current. Charge current: • All energy in a net/gates is wasted each time it toggles. • The energy in a capacitor is E = CV 2/2. • Dominant capacitance is proportional to net length. -
Analysis of Clocked Timing Elements for Dynamic Voltage Scaling Effects Over Process Parameter Variation Hoang Q
Analysis of Clocked Timing Elements for Dynamic Voltage Scaling Effects over Process Parameter Variation Hoang Q. Dao Kevin Nowka Vojin G. Oklobdzija ACSEL Lab IBM Austin Research Lab ACSEL Lab University of California, Davis 11400 Burnet Road MS9460, University of California, Davis 1 Shields Avenue, Davis, CA 95616 Austin, TX 78758 1 Shields Avenue, Davis, CA 95616 +1-530-754-6827 +1-512-838-3350 +1-530-754-6827 [email protected] [email protected] [email protected] ABSTRACT It was our interest to analyze the effect of large supply voltage In power-constrained systems, the power efficiency of latches and variation as well as the typical process and environmental flip-flops is pivotal. Characteristics of three selected latches and variation in bulk CMOS circuits. In assessing the quality of FFs were analyzed for their behavior under voltage scaling and designs for the low power, both the designs and the evaluation different process corners in a 0.18um CMOS technology. The methodology and criteria are equally important. We specifically relative performance amongst the latches/FFs was consistent focus on latches and flip-flops: an improved semi-dynamic flip- across the different supply voltages. At low-voltage power-delay- flop (SDFF) [1], the sense-amplifier flip-flop (SAFF) [2] and the product was degraded by about 25%. Energy-delay-product was master-slave PowerPC latch [5]. Section 2 presents the effects of approximately doubled at low-voltage – for all latches/FFs over voltage scaling. Section 3 discusses modification to the test all process corners. This result was smaller in comparison to the bench by Stojanovic and Oklobdzija [3]. -
Amd Power Management Srilatha Manne 11/14/2013 Outline
AMD POWER MANAGEMENT SRILATHA MANNE 11/14/2013 OUTLINE APUs and HSA Richland APU APU Power Management Monitoring and Management Features 2 | DOE WEBINAR| NOV 14, 2013 APU: ACCELERATED PROCESSING UNIT The APU has arrived and it is a great advance over previous platforms Combines scalar processing on CPU with parallel processing on the GPU and high-bandwidth access to memory Advantages over other forms of compute offload: Easier to program Easier to optimize Easier to load-balance Higher performance Lower power © Copyright 2012 HSA Foundation. All Rights Reserved. 3 WHAT IS HSA? Joins CPUs, GPUs, and accelerators into a unified computing framework Single address space accessible to avoid the overhead of data copying HSA Accelerated Applications Access to Broad Set of Programming Use-space queuing to minimize Languages communication overhead Pre-emptive context switching for better Legacy quality of service Applications Simplified programming HSA Runtime Single, standard computing environments Infrastructure Support for mainstream languages— OS C, C++, Fortran, Java, .NET Lower development costs Memory Optimized compute density Radical performance improvement for CPU GPU ACC HPC, big data, and multimedia workloads HSA Platform Low power to maximize performance per watt 4 | AMD OPTERON™ PROCESSOR ROADMAP | NOVEMBER 13, 2013 A NEW ERA OF PROCESSOR PERFORMANCE Heterogeneous Single-core Era Multi-core Era Systems Era Temporarily Enabled by: Constrained by: Enabled by: Constrained by: Enabled by: Moore’s Power Moore’s Law Power Abundant data Constrained by: Law Complexity SMP Parallel SW parallelism Programming Voltage architecture Scalability Power efficient models Scaling GPUs Comm.overhead pthreads OpenMP / TBB … Assembly C/C++ Java Shader CUDA OpenCL … C++ and Java ? we are here we are Performance Performance here Throughput Performance Performance we are Modern Modern Application here Single-thread Single-thread Performance Time Time (# of processors) Time (Data-parallel exploitation) © Copyright 2012 HSA Foundation.