Fast Cache Hit by Virtual Cache Fast Cache Hits by Avoiding Translation

Total Page:16

File Type:pdf, Size:1020Kb

Fast Cache Hit by Virtual Cache Fast Cache Hits by Avoiding Translation Fast Cache Hit by Virtual Cache Physical Cache – physically indexed and physically tagged cache Virtual Cache – virtually indexed and virtually tagged cache Lecture 18: VLIW and EPIC Must flush cache at process switch, or add PID Must handle virtual address alias to identical physical address V-addr Virtual Static superscalar, VLIW, EPIC and V-addr Physical cache Itanium Processor TLB cache P-index (First introduce fast and high- P-addr bandwidth L1 cache design) P-tag P-index TLB V-tag data PID P-addr P-tag data To L2 =? =? 1 2 Fast Cache Hits by Avoiding Virtually Indexed, Physically Tagged Translation: Process ID impact Cache What motivation? Black is uniprocess Fast cache hit by V-addr Light Gray is multiprocess parallel TLB access V-index No virtual cache when flush cache shortcomings TLB P-tag data Dark Gray is multiprocess P-addr when use Process ID tag How could it be correct? =? Y axis: Miss Rates up to Require cache way size 20% <= page size; now physical index is from X axis: Cache size from 2 page offset What if want bigger caches? KB to 1024 KB Then virtual and physical indices are Higher associativity identical ⇒ works like a moves barrier to right physically indexed Page coloring cache! 3 4 Pipelined Cache Access Pipelined Cache Access For multi-issue, cache bandwidth affects Alpha 21264 Data cache design effective cache hit time Queueing delay adds up if cache does not have The cache is 64KB, 2-way associative; enough read/write ports cannot be accessed within one-cycle Pipelined cache accesses: reduce cache cycle One-cycle used for address transfer and time and improve bandwidth data transfer, pipelined with data array Cache organization for high bandwidth access Duplicate cache Banked cache Cache clock frequency doubles processor Double clocked cache frequency; wave pipelined to achieve the Key technology: Wave pipelining that does not speed need latches 5 6 1 Trace Cache Two Paths to High ILP Trace: a dynamic sequence of instructions including taken branches Modern superscalar processors: dynamically scheduled, speculative execution, branch Traces are dynamically constructed by prediction, dynamic memory disambiguation, processor hardware and frequently used non-blocking cache => More and more traces are stored into trace cache hardware functionalities AND complexities Example: Intel P4 processor, storing about Another direction: Let complier take the 12K mops complexity Simple hardware, smart compiler (End of cache) Static Superscalar, VLIW, EPIC 7 8 MIPS Pipeline with pipelined multi-cycle More Hazards Detection and Forwarding Operations Assume checking hazards at ID (the simple way) Page A-50, Figure A.31 Structural hazards EX Hazards at WB: Track usages of registers at ID, stall instructions if hazards is detected Separate int and fp registers to reduce hazards M1 M2 M3 M4 M5 M6 M7 RAW hazards: Check source registers with all EX stages except the last ones. IF ID M WB A dependent instruction must wait for the producing A1 A2 A3 A4 instruction to reach the last stage of EX Ex: check with ID/A1, A1/A2, A2/A3, but not A4/MEM. WAW hazards DIV Instructions reach WB out-of-order check with all multi-cycle stages (A1-A4, D, M1-M7) for the same dest register Pipelined implementations ex: 7 outstanding MUL, 4 outstanding Out-of-order completion complicates the Add, unpipelined DIV. maintenance of precise exception In-order execution, out-of-order completion Tomasulo w/o ROB: out-of-order execution, out-of-order More forwarding data paths completion, in-order commit 9 10 Complier Optimization Loop unrolling Example: add a scalar to a vector: 1 cycle stall for (i=1000; i>0; i=i–1) 1 Loop:L.D F0,0(R1) 2 cycles stall x[i] = x[i] + s; 2 ADD.D F4,F0,F2 3 S.D 0(R1),F4 ;drop DSUBUI & BNEZ 4L.DF6,-8(R1) MIPS code 5 ADD.D F8,F6,F2 6S.D-8(R1),F8 ;drop DSUBUI & BNEZ Loop:L.D F0,0(R1) ;F0=vector element 7 L.D F10,-16(R1) stall for L.D, assume 1 cycles 8 ADD.D F12,F10,F2 9S.D-16(R1),F12 ;drop DSUBUI & BNEZ ADD.D F4,F0,F2 ;add scalar from F2 10 L.D F14,-24(R1) stall for ADD, assume 2 cycles 11 ADD.D F16,F14,F2 S.D 0(R1),F4 ;store result 12 S.D -24(R1),F16 13 DSUBUI R1,R1,#32 ;alter to 4*8 DSUBUI R1,R1,8 ;decrement pointer 14 BNEZ R1,LOOP BNEZ R1,Loop ;branch R1!=zero 15 NOP stall for taken branch, assume 1 cycle 11 12 2 Unrolled Loop That Minimizes Stalls Register Renaming 1 Loop:L.D F0,0(R1) 1 Loop:L.D F0,0(R1) 1 Loop:L.D F0,0(R1) 2 ADD.D F4,F0,F2 2 ADD.D F4,F0,F2 2 L.D F6,-8(R1) Called code 3 S.D 0(R1),F4 3 S.D 0(R1),F4 3 L.D F10,-16(R1) 4L.DF0,-8(R1) 4L.DF6,-8(R1) 4 L.D F14,-24(R1) movement 5 ADD.D F4,F0,F2 5 ADD.D F8,F6,F2 5 ADD.D F4,F0,F2 6S.D-8(R1),F4 6S.D-8(R1),F8 6 ADD.D F8,F6,F2 Moving store past 7L.DF0,-16(R1) 7L.DF10,-16(R1) 7 ADD.D F12,F10,F2 DSUBUI 8 ADD.D F4,F0,F2 8 ADD.D F12,F10,F2 8 ADD.D F16,F14,F2 Moving loads before 9S.D-16(R1),F4 9S.D-16(R1),F12 9 S.D 0(R1),F4 10 L.D F0,-24(R1) 10 L.D F14,-24(R1) 10 S.D -8(R1),F8 stores 11 ADD.D F4,F0,F2 11 ADD.D F16,F14,F2 11 S.D -16(R1),F12 12 S.D -24(R1),F4 12 S.D -24(R1),F16 12 DSUBUI R1,R1,#32 13 DSUBUI R1,R1,#32 13 DSUBUI R1,R1,#32 13 BNEZ R1,LOOP 14 BNEZ R1,LOOP 14 BNEZ R1,LOOP 14 S.D 8(R1),F16 ; delayed branch slot 15 NOP 15 NOP Original register renaming 13 14 VLIW: Very Large Instruction Word VLIW Example: Loop Unrolling Memory Memory FP FP Int. op/ Clock Static Superscalar: hardware detects hazard, complier reference 1 reference 2 operation 1 op. 2 branch determines scheduling L.D F0,0(R1) L.D F6,-8(R1) 1 VLIW: complier takes both jobs L.D F10,-16(R1) L.D F14,-24(R1) 2 L.D F18,-32(R1) L.D F22,-40(R1) ADD.D F4,F0,F2 ADD.D F8,F6,F2 3 Each “instruction” has explicit coding for multiple L.D F26,-48(R1) ADD.D F12,F10,F2 ADD.D F16,F14,F2 4 operations ADD.D F20,F18,F2 ADD.D F24,F22,F2 5 S.D 0(R1),F4 S.D -8(R1),F8 ADD.D F28,F26,F2 6 S.D -16(R1),F12 S.D -24(R1),F16 7 There is no or only partial hardware hazard detection S.D -32(R1),F20 S.D -40(R1),F24 DSUBUI R1,R1,#48 8 No dependence check logic for instruction issued at the same S.D -0(R1),F28 BNEZ R1,LOOP 9 cycle Wide instruction format allows theoretically high ILP Unrolled 7 times to avoid delays 7 results in 9 clocks, or 1.3 clocks per iteration (1.8X) Tradeoff instruction space for simple decoding Average: 2.5 ops per clock, 50% efficiency The long instruction word has room for many operations But have to fill with NOOP if no enough operations are found Note: Need more registers in VLIW (15 in this example) 15 16 Problems with First Generation VLIW Scheduling Across Branches Increase in code size Local scheduling or basic block scheduling Wasted issue slots are translated to no-ops in instruction encoding Typically in a range of 5 to 20 instructions About 50% instructions are no-ops Unrolling may increase basic block size to facilitate The increase is from 5 instructions to 45 scheduling instructions! However, what happens if branches exist in loop body? Operated in lock-step; no hazard detection HW Any function unit stalls → The entire processor stalls Global scheduling: moving instructions across Compiler can schedule around function unit stalls, but branches (i.e., cross basic blocks) how about cache miss? We cannot change data flow with any branch outputs Compilers are not allowed to speculate! How to guarantee correctness? Binary code compatibility Increase scheduling scope: trace scheduling, Re-compile if #FU or any FU latency changes; not a superblock, predicted execution, etc. major problem today 17 18 3 EPIC/ IA-64: Motivation in 1989 EPIC, IA-64, and Itanium “First, it was quite evident from Moore's law that it would soon be possible to fit an entire, highly EPIC: Explicit Parallel Instruction parallel, ILP processor on a chip. Computing, an architecture framework Second, we believed that the ever-increasing complexity of superscalar processors would have a proposed by HP negative impact upon their clock rate, eventually leading to a leveling off of the rate of increase in microprocessor performance.” IA-64: An architecture that HP and Intel Schlansker and Rau, Computer Feb. 2000 developed under the EPIC framework Obvious today: Think about the complexity of P4, 21264, and other superscalar processor; processor complexity has been discussed in many Itanium: The first commercial processor papers since mid-1990s that implements IA-64 architecture; now Agarwal et al, "Clock rate versus IPC: The end of the road for conventional microarchitectures," ISCA Itanium 2 2000 19 20 Example: IF-conversion EPIC Main ideas Example of Itanium Compile does the scheduling code: Permitting the compiler to play the cmp.eq p1, p2 = rl, r2;; statistics (profiling) (p1) sub r9 = r10, r11 Hardware supports speculation (p2) add r5 = r6, r7 Addressing the branch problem: predicted execution and many other techniques Addressing the memory problem: cache specifiers, prefetching, speculation on memory alias Use of predicated execution to perform if- conversion: Eliminate the branch and produces just one basic block containing operations guarded by the appropriate predicates.
Recommended publications
  • Chapter 1: Computer Abstractions and Technology 1.6 – 1.7: Performance and Power
    Chapter 1: Computer Abstractions and Technology 1.6 – 1.7: Performance and power ITSC 3181 Introduction to Computer Architecture https://passlaB.githuB.io/ITSC3181/ Department of Computer Science Yonghong Yan [email protected] https://passlab.github.io/yanyh/ Lectures for Chapter 1 and C Basics Computer Abstractions and Technology • Lecture 01: Chapter 1 – 1.1 – 1.4: Introduction, great ideas, Moore’s law, aBstraction, computer components, and program execution • Lecture 02: C Basics; Memory and Binary Systems • Lecture 03: Number System, Compilation, Assembly, Linking and Program Execution ☛• Lecture 04: Chapter 1 – 1.6 – 1.7: Performance, power and technology trends • Lecture 05: – 1.8 - 1.9: Multiprocessing and Benchmarking 2 § 1.6 Performance 1.6 Defining Performance • Which airplane has the best performance? Boeing 777 Boeing 777 Boeing 747 Boeing 747 BAC/Sud BAC/Sud Concorde Concorde Douglas Douglas DC- DC-8-50 8-50 0 100 200 300 400 500 0 2000 4000 6000 8000 10000 Passenger Capacity Cruising Range (miles) Boeing 777 Boeing 777 Boeing 747 Boeing 747 BAC/Sud BAC/Sud Concorde Concorde Douglas Douglas DC- DC-8-50 8-50 0 500 1000 1500 0 100000 200000 300000 400000 Cruising Speed (mph) Passengers x mph 3 Response Time and Throughput • Response time çè Latency – How long it takes to do a task • Throughput çè Bandwidth – Total work done per unit time • e.g., tasks/transactions/… per hour • How are response time and throughput affected by – Replacing the processor with a faster version? – Adding more processors? • We’ll focus on response time for now… 4 Relative Performance • Define Performance = 1/Execution Time • “X is n time faster than Y”, i.e.
    [Show full text]
  • Clock Rate Improves Roughly Proportional to Improvement in L • Number of Transistors Improves Proportional to L2 (Or Faster)
    TheThe VonVon NeumannNeumann ComputerComputer ModelModel • Partitioning of the computing engine into components: – Central Processing Unit (CPU): Control Unit (instruction decode , sequencing of operations), Datapath (registers, arithmetic and logic unit, buses). – Memory: Instruction and operand storage. – Input/Output (I/O) sub-system: I/O bus, interfaces, devices. – The stored program concept: Instructions from an instruction set are fetched from a common memory and executed one at a time Control Input Memory - (instructions, data) Datapath registers Output ALU, buses Computer System CPU I/O Devices EECC551 - Shaaban #1 Lec # 1 Winter 2001 12-3-2001 Generic CPU Machine Instruction Execution Steps Instruction Obtain instruction from program storage Fetch Instruction Determine required actions and instruction size Decode Operand Locate and obtain operand data Fetch Execute Compute result value or status Result Deposit results in storage for later use Store Next Determine successor or next instruction Instruction EECC551 - Shaaban #2 Lec # 1 Winter 2001 12-3-2001 HardwareHardware ComponentsComponents ofof AnyAny ComputerComputer Five classic components of all computers: 1. Control Unit; 2. Datapath; 3. Memory; 4. Input; 5. Output } Processor Computer Keyboard, Mouse, etc. Processor Memory Devices (active) (passive) Control Input (where Unit programs, data Disk Datapath live when Output running) Display, Printer, etc. EECC551 - Shaaban #3 Lec # 1 Winter 2001 12-3-2001 CPUCPU OrganizationOrganization • Datapath Design: – Capabilities & performance characteristics of principal Functional Units (FUs): • (e.g., Registers, ALU, Shifters, Logic Units, ...) – Ways in which these components are interconnected (buses connections, multiplexors, etc.). – How information flows between components. • Control Unit Design: – Logic and means by which such information flow is controlled. – Control and coordination of FUs operation to realize the targeted Instruction Set Architecture to be implemented (can either be implemented using a finite state machine or a microprogram).
    [Show full text]
  • Atmega165p Datasheet
    Features • High Performance, Low Power Atmel® AVR® 8-Bit Microcontroller • Advanced RISC Architecture – 130 Powerful Instructions – Most Single Clock Cycle Execution – 32 × 8 General Purpose Working Registers – Fully Static Operation – Up to 16 MIPS Throughput at 16 MHz – On-Chip 2-cycle Multiplier • High Endurance Non-volatile Memory segments – 16 Kbytes of In-System Self-programmable Flash program memory – 512 Bytes EEPROM – 1 Kbytes Internal SRAM 8-bit – Write/Erase cyles: 10,000 Flash/100,000 EEPROM(1)(3) – Data retention: 20 years at 85°C/100 years at 25°C(2)(3) Microcontroller – Optional Boot Code Section with Independent Lock Bits In-System Programming by On-chip Boot Program with 16K Bytes True Read-While-Write Operation – Programming Lock for Software Security In-System • JTAG (IEEE std. 1149.1 compliant) Interface – Boundary-scan Capabilities According to the JTAG Standard Programmable – Extensive On-chip Debug Support – Programming of Flash, EEPROM, Fuses, and Lock Bits through the JTAG Interface Flash • Peripheral Features – Two 8-bit Timer/Counters with Separate Prescaler and Compare Mode – One 16-bit Timer/Counter with Separate Prescaler, Compare Mode, and Capture Mode – Real Time Counter with Separate Oscillator –Four PWM Channels ATmega165P – 8-channel, 10-bit ADC – Programmable Serial USART ATmega165PV – Master/Slave SPI Serial Interface – Universal Serial Interface with Start Condition Detector – Programmable Watchdog Timer with Separate On-chip Oscillator – On-chip Analog Comparator Preliminary – Interrupt and Wake-up
    [Show full text]
  • Performance of a Computer (Chapter 4) Vishwani D
    ELEC 5200-001/6200-001 Computer Architecture and Design Fall 2013 Performance of a Computer (Chapter 4) Vishwani D. Agrawal & Victor P. Nelson epartment of Electrical and Computer Engineering Auburn University, Auburn, AL 36849 ELEC 5200-001/6200-001 Performance Fall 2013 . Lecture 1 What is Performance? Response time: the time between the start and completion of a task. Throughput: the total amount of work done in a given time. Some performance measures: MIPS (million instructions per second). MFLOPS (million floating point operations per second), also GFLOPS, TFLOPS (1012), etc. SPEC (System Performance Evaluation Corporation) benchmarks. LINPACK benchmarks, floating point computing, used for supercomputers. Synthetic benchmarks. ELEC 5200-001/6200-001 Performance Fall 2013 . Lecture 2 Small and Large Numbers Small Large 10-3 milli m 103 kilo k 10-6 micro μ 106 mega M 10-9 nano n 109 giga G 10-12 pico p 1012 tera T 10-15 femto f 1015 peta P 10-18 atto 1018 exa 10-21 zepto 1021 zetta 10-24 yocto 1024 yotta ELEC 5200-001/6200-001 Performance Fall 2013 . Lecture 3 Computer Memory Size Number bits bytes 210 1,024 K Kb KB 220 1,048,576 M Mb MB 230 1,073,741,824 G Gb GB 240 1,099,511,627,776 T Tb TB ELEC 5200-001/6200-001 Performance Fall 2013 . Lecture 4 Units for Measuring Performance Time in seconds (s), microseconds (μs), nanoseconds (ns), or picoseconds (ps). Clock cycle Period of the hardware clock Example: one clock cycle means 1 nanosecond for a 1GHz clock frequency (or 1GHz clock rate) CPU time = (CPU clock cycles)/(clock rate) Cycles per instruction (CPI): average number of clock cycles used to execute a computer instruction.
    [Show full text]
  • Chap01: Computer Abstractions and Technology
    CHAPTER 1 Computer Abstractions and Technology 1.1 Introduction 3 1.2 Eight Great Ideas in Computer Architecture 11 1.3 Below Your Program 13 1.4 Under the Covers 16 1.5 Technologies for Building Processors and Memory 24 1.6 Performance 28 1.7 The Power Wall 40 1.8 The Sea Change: The Switch from Uniprocessors to Multiprocessors 43 1.9 Real Stuff: Benchmarking the Intel Core i7 46 1.10 Fallacies and Pitfalls 49 1.11 Concluding Remarks 52 1.12 Historical Perspective and Further Reading 54 1.13 Exercises 54 CMPS290 Class Notes (Chap01) Page 1 / 24 by Kuo-pao Yang 1.1 Introduction 3 Modern computer technology requires professionals of every computing specialty to understand both hardware and software. Classes of Computing Applications and Their Characteristics Personal computers o A computer designed for use by an individual, usually incorporating a graphics display, a keyboard, and a mouse. o Personal computers emphasize delivery of good performance to single users at low cost and usually execute third-party software. o This class of computing drove the evolution of many computing technologies, which is only about 35 years old! Server computers o A computer used for running larger programs for multiple users, often simultaneously, and typically accessed only via a network. o Servers are built from the same basic technology as desktop computers, but provide for greater computing, storage, and input/output capacity. Supercomputers o A class of computers with the highest performance and cost o Supercomputers consist of tens of thousands of processors and many terabytes of memory, and cost tens to hundreds of millions of dollars.
    [Show full text]
  • RAMP: Research Accelerator for Multiple Processors
    Technical Report UCB//CSD-05-1412, September 2005 RAMP: Research Accelerator for Multiple Processors - A Community Vision for a Shared Experimental Parallel HW/SW Platform Arvind (MIT), Krste Asanovic´ (MIT), Derek Chiou (UT Austin), James C. Hoe (CMU), Christoforos Kozyrakis (Stanford), Shih-Lien Lu (Intel), Mark Oskin (U Washington), David Patterson (UC Berkeley), Jan Rabaey (UC Berkeley), and John Wawrzynek (UC Berkeley) Project Summary Desktop processor architectures have crossed a critical threshold. Manufactures have given up attempting to extract ever more performance from a single core and instead have turned to multi-core designs. While straightforward approaches to the architecture of multi-core processors are sufficient for small designs (2–4 cores), little is really known how to build, program, or manage systems of 64 to 1024 processors. Unfortunately, the computer architecture community lacks the basic infrastructure tools required to carry out this research. While simulation has been adequate for single-processor research, significant use of simplified modeling and statistical sampling is required to work in the 2–16 processing core space. Invention is required for architecture research at the level of 64–1024 cores. Fortunately, Moore’s law has not only enabled these dense multi-core chips, it has also enabled extremely dense FPGAs. Today, for a few hundred dollars, undergraduates can work with an FPGA prototype board with almost as many gates as a Pentium. Given the right support, the research community can capitalize on this opportunity too. Today, one to two dozen cores can be programmed into a single FPGA. With multiple FPGAs on a board and multiple boards in a system, large complex architectures can be explored.
    [Show full text]
  • Computer “Performance”
    Computer “Performance” Readings: 1.6-1.8 BIPS (Billion Instructions Per Second) vs. GHz (Giga Cycles Per Second) Throughput (jobs/seconds) vs. Latency (time to complete a job) Measuring “best” in a computer Hyper 3.0 GHz The PowerBook G4 outguns Pentium Pipelined III-based notebooks by up to 30 percent.* Technology * Based on Adobe Photoshop tests comparing a 500MHz PowerBook G4 to 850MHz Pentium III-based portable computers 58 Performance Example: Homebuilders Builder Time per Houses Per House Dollars Per House Month Options House Self-build 24 months 1/24 Infinite $200,000 Contractor 3 months 1 100 $400,000 Prefab 6 months 1,000 1 $250,000 Which is the “best” home builder? Homeowner on a budget? Rebuilding Haiti? Moving to wilds of Alaska? Which is the “speediest” builder? Latency: how fast is one house built? Throughput: how long will it take to build a large number of houses? 59 Computer Performance Primary goal: execution time (time from program start to program completion) 1 Performance ExecutionTime To compare machines, we say “X is n times faster than Y” Performance ExecutionTime n x y Performancey ExecutionTimex Example: Machine Orange and Grape run a program Orange takes 5 seconds, Grape takes 10 seconds Orange is _____ times faster than Grape 60 Execution Time Elapsed Time counts everything (disk and memory accesses, I/O , etc.) a useful number, but often not good for comparison purposes CPU time doesn't count I/O or time spent running other programs can be broken up into system time, and user time Example: Unix “time” command linux15.ee.washington.edu> time javac CircuitViewer.java 3.370u 0.570s 0:12.44 31.6% Our focus: user CPU time time spent executing the lines of code that are "in" our program 61 CPU Time CPU execution time CPU clock cycles =*Clock period for a program for a program CPU execution time CPU clock cycles 1 =* for a program for a program Clock rate Application example: A program takes 10 seconds on computer Orange, with a 400MHz clock.
    [Show full text]
  • Intel® Itanium® Architecture Software Developer's Manual
    Intel® Itanium® Architecture Software Developer’s Manual Volume 2: System Architecture Revision 2.1 October 2002 Document Number: 245318-004 THIS DOCUMENT IS PROVIDED “AS IS” WITH NO WARRANTIES WHATSOEVER, INCLUDING ANY WARRANTY OF MERCHANTABILITY, FITNESS FOR ANY PARTICULAR PURPOSE, OR ANY WARRANTY OTHERWISE ARISING OUT OF ANY PROPOSAL, SPECIFICATION OR SAMPLE. INFORMATION IN THIS DOCUMENT IS PROVIDED IN CONNECTION WITH INTEL® PRODUCTS. NO LICENSE, EXPRESS OR IMPLIED, BY ESTOPPEL OR OTHERWISE, TO ANY INTELLECTUAL PROPERTY RIGHTS IS GRANTED BY THIS DOCUMENT. EXCEPT AS PROVIDED IN INTEL'S TERMS AND CONDITIONS OF SALE FOR SUCH PRODUCTS, INTEL ASSUMES NO LIABILITY WHATSOEVER, AND INTEL DISCLAIMS ANY EXPRESS OR IMPLIED WARRANTY, RELATING TO SALE AND/OR USE OF INTEL PRODUCTS INCLUDING LIABILITY OR WARRANTIES RELATING TO FITNESS FOR A PARTICULAR PURPOSE, MERCHANTABILITY, OR INFRINGEMENT OF ANY PATENT, COPYRIGHT OR OTHER INTELLECTUAL PROPERTY RIGHT. INTEL PRODUCTS ARE NOT INTENDED FOR USE IN MEDICAL, LIFE SAVING, OR LIFE SUSTAINING APPLICATIONS. Intel may make changes to specifications and product descriptions at any time, without notice. Designers must not rely on the absence or characteristics of any features or instructions marked "reserved" or "undefined." Intel reserves these for future definition and shall have no responsibility whatsoever for conflicts or incompatibilities arising from future changes to them. Intel processors based on the Itanium architecture may contain design defects or errors known as errata which may cause the product to deviate from published specifications. Current characterized errata are available on request. Contact your local Intel sales office or your distributor to obtain the latest specifications and before placing your product order.
    [Show full text]
  • The Role of Performance Relating the Metrics CPU Execution Time for a Program = CPU Clock Cycles for a Program * Ming-Hwa Wang, Ph.D
    The Role of Performance Relating the Metrics CPU execution time for a program = CPU clock cycles for a program * Ming-Hwa Wang, Ph.D. clock cycle time = CPU clock cycles for a program / clock rate, measured COEN 210 Computer Architecture by running the program Department of Computer Engineering CPU clock cycles = instructions for a program * average CPI or clock Santa Clara University cycles per instruction measuring the instruction count (depends on the architecture but not on the Introduction exact implementation) by using hardware counters or by using software hardware performance is often key to the effectiveness of an entire system tools (that profile the execution or by using a simulator of the architecture) of hardware and software CPI = CPU clock cycles / instruction count, provides one way of comparing accurately measuring and comparing different machines is critical to two different implementations of the same instruction set architecture, CPI purchasers, and therefore to designers varies by applications as well as among implementations with the same for different types of applications, different performance metrics may be instruction set appropriate and different aspects of a computer system may be the most CPU time = instruction count * CPI * clock cycle time = instruction count * significant in determining overall performance CPI / clock rate an individual computer user is interested in reducing response/execution CPU execution time = (instructions / program) * (clock cycles / instruction) * time (the time between the start
    [Show full text]
  • Chapter 4 Accessing and Understanding Performance
    ChapterChapter 4 4 AccessingAccessing and and Understanding Understanding PerformancePerformance 1 Performance Why do we care about performance evaluation? –Purchasing perspective given a collection of machines, which has the –best performance ? –least cost ? –best performance / cost ? –Design perspective faced with design options, which has the –best performance improvement ? –least cost ? –best performance / cost ? How to measure, report, and summarize performance? –Performance metric –Benchmark 2 Which of these airplanes has the best performance? What metric defines performance? –Capacity, cruising range, or speed? Speed –Taking one passenger from one point to another in the least time –Transporting 450 passengers from one point to another 3 Two Notions of “Performance” Response Time (latency) –How long does it take for my job to run? –How long does it take to execute a job? –How long must I wait for the database query? Throughput –How many jobs can the machine run at once? –What is the average execution rate? –How much work is getting done? If we upgrade a machine with a new processor what do we increase? If we add a new machine to the lab what do we increase? 4 Execution Time Elapsed Time –counts everything (disk and memory accesses, I/O , etc.) –a useful number, but often not good for comparison purposes CPU time –doesn't count I/O or time spent running other programs –can be broken up into system time, and user time Our focus: user CPU time –time spent executing the lines of code that are "in" our program 5 Definitions Performance is in units of things-per-second –bigger is better If we are primarily concerned with response time " X is n times faster than Y" means 6 Which one is faster? Concorde or Boeing 747 Response Time of Concorde vs.
    [Show full text]
  • TMS320F28035-EP Piccolo™ Microcontroller
    Product Order Technical Tools & Support & Folder Now Documents Software Community TMS320F28035-EP SPRSP25A –JUNE 2018–REVISED JULY 2018 TMS320F28035-EP Piccolo™ Microcontroller 1 Device Overview 1.1 Features 1 • High-Efficiency 32-Bit CPU (TMS320C28x) • Code-Security Module – 60 MHz (16.67-ns Cycle Time) • 128-Bit Security Key and Lock – 16 × 16 and 32 × 32 MAC Operations – Protects Secure Memory Blocks – 16 × 16 Dual MAC – Prevents Firmware Reverse Engineering – Harvard Bus Architecture • Serial Port Peripherals – Atomic Operations – One Serial Communications Interface (SCI) – Fast Interrupt Response and Processing Universal Asynchronous Receiver/Transmitter – Unified Memory Programming Model (UART) Module – Code-Efficient (in C/C++ and Assembly) – Two Serial Peripheral Interface (SPI) Modules • Programmable Control Law Accelerator (CLA) – One Inter-Integrated-Circuit (I2C) Module – 32-Bit Floating-Point Math Accelerator – One Local Interconnect Network (LIN) Module – Executes Code Independently of the Main CPU – One Enhanced Controller Area Network (eCAN) Module • Endianness: Little Endian • Enhanced Control Peripherals • JTAG Boundary Scan Support – ePWM – IEEE Standard 1149.1-1990 Standard Test Access Port and Boundary Scan Architecture – High-Resolution PWM (HRPWM) • Low Cost for Both Device and System: – Enhanced Capture (eCAP) Module – Single 3.3-V Supply – High-Resolution Input Capture (HRCAP) Module – No Power Sequencing Requirement – Enhanced Quadrature Encoder Pulse (eQEP) Module – Integrated Power-On Reset and Brown-Out Reset
    [Show full text]
  • Introduction to Microprocessors
    Introduction to Microprocessors Yuri Baida [email protected] [email protected] October 2, 2010 MDSP Project | Intel Lab Moscow Institute of Physics and Technology Agenda • Background and History – What is a microprocessor? – What is the history of the development of the microprocessor? – How does transistor scaling affect processor design? • PC Components – What are the major PC components and their functions? – What is memory hierarchy and how has it changed? • Processor Architecture – What are processor architecture and microarchitecture? – How does microarchitecture affect performance? – How is performance measured? MDSP Project | Intel Lab Moscow Institute of Physics and Technology 2 Background and History MDSP Project | Intel Lab Moscow Institute of Physics and Technology 3 What is a Microprocessor? • Microprocessor is a computer Central Processing Unit (CPU) on a single chip. • It contains millions of transistors connected by wires Core i7 die Core i7 in package Picture: Intel Picture: Ebbesen MDSP Project | Intel Lab Moscow Institute of Physics and Technology 4 Electrical Numerical Integrator and Calculator • Designed for the U.S. Army's Ballistic Research Laboratory • Built out of – 17,468 vacuum tubes – 7,200 crystal diodes – 1,500 relays – 70,000 resistors – 10,000 capacitors • Consumed 150 kW of power • Took up 72 m2 • Weighted 27 tons • Suffered a failure on average every 6 hours MDSP Project | Intel Lab Moscow Institute of Physics and Technology 5 Electrical Numerical Integrator and Calculator Glen Beck and Betty
    [Show full text]