CS 6534: Tech Trends / Intro

CS 6534: Tech Trends / Intro

CS 6534: Tech Trends / Intro Charles Reiss 24 August 2016 1 Moore’s Law Microprocessor Transistor Counts 1971-2011 & Moore's Law 16-Core SPARC T3 Six-Core Core i7 2,600,000,000 Six-Core Xeon 7400 10-Core Xeon Westmere-EX Dual-Core Itanium 2 8-core POWER7 Quad-core z196 AMD K10 Quad-Core Itanium Tukwila 1,000,000,000 POWER6 8-Core Xeon Nehalem-EX Itanium 2 with 9MB cache Six-Core Opteron 2400 AMD K10 Core i7 (Quad) Core 2 Duo Itanium 2 Cell 100,000,000 AMD K8 Barton Pentium 4 Atom AMD K7 AMD K6-III AMD K6 10,000,000 Pentium III Pentium II AMD K5 Pentium 1,000,000 80486 80386 Transistor count Transistor 80286 100,000 68000 80186 8086 8088 8085 10,000 6800 6809 8080 Z80 8008 MOS 6502 2,300 4004 RCA 1802 1971 1980 1990 2000 2011 Date of introduction Wikimedia Commons / Wgsimon 2 Good Ol’ Days: Frequency Scaling Figure 1.11 Growth in clock rate of microprocessors in Figure 1.1. Between 1978 and 1986, the clock rate improved less than 15% per year while performance improved by 25% per year. During the “renaissance period” of 52% performance improvement per year between 1986 and 2003, clock rates shot up almost 40% per year. Since then, the clock rate has been nearly flat, growing at less than 1% per year, while single processor performance improved at less than 22% per year. H&P 3 Copyright © 2011, Elsevier Inc. All rights Reserved. 5 The Power Wall Power ∼ Switching Power + Leakage Power Switching Power ∼ Capacitance×Voltage2×Frequency 4 Increasing Parallelism: Cores 25 20 15 10 5 0 200520062008200920102012201320142016 Intel x86 # of Cores Per Package Date 5 Increasing Parallelism: Vector width 500 x86 ARM 400 300 200 100 vector register size (bits) 0 1995 1997 2000 2003 2005 2008 2011 2014 2016 Date 6 Increasing Parallelism: ILP x86 Intel 32-bit adds per cycle 4 3 2 1 0 1978 1983 1988 1994 1999 2005 2010 2016 Date 7 Limits: Parallelism Amdahl’s Law 20 15 5% serial 0% serial 10 10% serial 5 25% serial Speedup (1=serial) 50% serial 10 20 30 40 50 60 Degree of Parallelism (1=serial) 8 Limits: Communication [Balfour et al, “Operand Registers and Explicit Operand Forwarding”, 2009.] [Malladi et al, “Towards Energy-Proportional Datacenter Memory with Mobile DRAM”, 2012.] DDR3 DRAM (32-bit read/write) full utilization 2 300 000 fJ 4300× low utilization 7 700 000 fJ 15000× 9 Increasing Efficiency: Specialization Task/workload-specific coprocessors or instructions Maybe reconfigurable? Heterogeneous systems different parts for different types of computation 10 Interlude: Logistics Paper reviews — approx 2/class Homeworks — programming assignments Exam — end of semester 11 Textbook? Primarily paper readings Classic + some newish papers Reference: Hennessy and Patterson, Computer Architecture: A Quantitative Approach 12 Might not be what the authors put in their abstract/conclusion Paper Reviews What was your most significant insight from the paper? What evidence does the paper have to support this insight? What is the weakest part of the paper or how could it be approved? What topic from the paper would you like to see discussed in class (if any)? 13 Paper Reviews What was your most significant insight from the paper? Might not be what the authors put in their abstract/conclusion What evidence does the paper have to support this insight? What is the weakest part of the paper or how could it be approved? What topic from the paper would you like to see discussed in class (if any)? 13 Paper Discussions and not paper lectures. Requires your cooperation. 14 Homeworks Individual programming + writing assignments First — on memory hierarchy — available now. Second — to be announced — likely on superscalar Third — to be announced — likely GPU programming 15 Example: 32K cache means accessing 32K repeatedly is faster than 128K repeatedly. Homework 1 Description on course website (linked off Collab) Memory system parameters by benchmarking 16 Homework 1 Description on course website (linked off Collab) Memory system parameters by benchmarking Example: 32K cache means accessing 32K repeatedly is faster than 128K repeatedly. 16 Homework 1: Disclaimer This is probably hard Modern memory hierarchies are complicated Documentation is incomplete Mainly looking for: measurement technique that ‘should’ work If it doesn’t, try to come up with good reasons why 17 Exam There will be in final, probably in-class. Cover material from papers, homeworks, discussions in class 18 Exceptions / etc. Need accommodations — please ask Disability accommodations — Student Disability Access Center 19 Asking Questions Piazza (linked of Collab) Office Hours: Instructor Lecturer Charles Reiss TA Luowan Wang Loation Soda 205 TBA Times Monday 1PM–3PM Tuesday 1PM–2PM Friday 10AM–noon Email: [email protected] 20 Survey linked off Collab anonymous please do it 21 Preview of coming topics 22 Memory hierarchy caching — review(?) and advanced techniques homework 1 23 Pipelining different parts of multiple instructions at the same time more advanced topics: handling exceptions Image: Wikimedia commons / Poil 24 Increasing Parallelism: ILP x86 Intel 32-bit adds per cycle 4 3 2 1 0 1978 1983 1988 1994 1999 2005 2010 2016 Date 25 Beyond pipelining: Multiple issue starting multiple instructions at the same time allows cycles per instruction < 1 26 Beyond pipelining: Out-of-order run next instruction despite stall of prior one slow cache read-after-write hazard ... speculation — guess outcome of branch/load/etc. fix later if wrong 27 Increasing Parallelism: Cores 25 20 15 10 5 0 200520062008200920102012201320142016 Intel x86 # of Cores Per Package Date 28 Multiprocessor/multicore connecting processors together shared memory — multiple threads synchronization 29 Increasing Parallelism: Vector width 500 x86 ARM 400 300 200 100 vector register size (bits) 0 1995 1997 2000 2003 2005 2008 2011 2014 2016 Date 30 Vector/SIMD/GPUs single instruction/multiple data started with early supercomputers basis of GPU programming model 31 Specialization using custom chips (or circuits within chips) reconfigurable processors (e.g. FPGAs) 32 Miscellaneous topics hardware security warehouse-scale computers . depends on time Suggestions? 33 Papers for Next Class Alan Smith’s review of caching in 1982 D. J. Bernstein’s timing attack and suggestions to computer architects in 2005 Note: We’re not reading this to learn about AES 34.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    36 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us