Hypertransport Consortium Globalpress Presentation

Total Page:16

File Type:pdf, Size:1020Kb

Hypertransport Consortium Globalpress Presentation HyperTransportTM Technology Tutorial Pict u r e Prof. José Duato Technical University of Valencia, Spain Simula Research Laboratory, Oslo, Norway HyperTransport Technology Consortium Hot Chips Symposium August 23, 2009 www.hypertransport.org Copyright HyperTransport Consortium, 2009 1 With us Today and Happy to AddressYour Questions Brian Holden Mario Cavalli VP and Chair, Technical Working Group General Manager HyperTransport Technology Consortium HyperTransport Technology Consortium [email protected] [email protected] 408-472-6310 925-968-0220 Topics: Topics: HyperTransport Technology HyperTransport Market Positioning HyperTransport Consortium Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 2 Topics • Scope and Design Goals • HyperTransport Defined • Host I nterface • Connecting Device to Host • Connecting Multiple Devices to Host • InterconnectingMultipleHosts • AMD Cache Coherence Support • Beyond Motherboards • New in HT3 • Beyond HT3 • Beyond Conventional • HyperTransport Technology Consortium Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 3 Scope and Design Goals • System Area Network Supporting Cache-Coherent Shared-Memory Multiprocessors and I / O Devices • High-Performance Replacement for Processor Front Side Bus (Point-to-Point Links vs. Bus) Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 4 Scope and Design Goals (cont.) • System Area Network Supports Cache-Coherent Shared-Memory Multiprocessors and I / O Devices • High-Performance Replacement for Processor Front Side Bus (Point-to-Point Links vs. Bus) • HyperTransport’s Distinction: Processor-Native • Integrated in Processor Architectures Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 5 Scope and Design Goals (cont.) • Lowest-Latency, High Bandwidth, Cost-Effective, Reliable Motherboard-Level I nterconnect • SMP Programming Model • Unified I nterface For Local and Remote Memory • Self-Configuring Topology and Link Speed • HT3 Enhancements • IncreasedBandwidthand Reliability • Link Splitting • Dynamic Power Management • AC Mode • Hot Plugging Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 6 Typical Server Architecture Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 7 HyperTransportTM Defined Nine Years of Fine Tuning, Perfecting, Polishing Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 8 Host I nterface Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 9 Host I nterface (cont.) • Single I nterface for All Cores ( SRQ/ SRI ) • On-Chip Crossbar and Routing • Host Bridge for I / O Device Chain Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 10 Northbridge Architecture Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 11 Connecting Device to Host • Approach: High Speed Point-to-Point Parallel Link • Point-to-Point Link Minimizes Parasitic Capacitance • Clock Forwarding Removes Clock Recovery Overhead • Parallel Link Delivers High Bandwidth, Low Latency • Control and Data Packets Interleaved on Each Link • CTL Signal Distinguishes Between Control and Data • Tw o Additional System Signals: PWROK and RESET Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 12 HT Physical Layer • Low-Voltage Differential Signalling (LVDS) • Pre-Emphasis Supports Higher Clock Rates Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 13 HT Physical Layer (cont.) • Clock Rate: From 200 MHz to 3.2 GHz • Link Width: 2, 4, 8, 16, 32-Bit Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 14 HT Physical Layer (cont.) • Support for Asymmetric and Mixed Link Width Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 15 HT Transaction/ Data Link Layer Header DATA 8 or 12 4-64 Bytes Bytes Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 16 HT Basic Read/ Write Sequences Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 17 HT Request Packet Format Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 18 HT Request Packet Format (cont.) Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 19 HT Request Packet Format (cont.) Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 20 HT Read Response Packet Format Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 21 Com m on HT Com m and Types Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 23 Connecting Multiple Devices to Host Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 24 HyperTransport I / O Device Configurations Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 25 Routing to Target Device Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 26 Pipelining Multiple Requests Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 27 Communication Between Two I/ O Devices Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 28 Communication Between Two I/ O Devices (cont.) Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 29 Transmission Error Handling Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 30 Priority Request InterleavingTM Data transfer 1 under way CPU Peripheral A Peripheral B 1 PRI While transfer 1 carries on… CPU Peripheral A Peripheral B 2 Data transfer 2 initiates while data transfer 1 still under way CPU Peripheral A Peripheral B 3 Lowest Achievable Latency Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 31 InterconnectingMultipleHosts • Coherent vs. Non-Coherent HyperTransport • cHT-Enabled Links Configurable at Boot Time Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 34 Cache Coherence Support Proprietary Technology Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 35 AMD cHT Basics • On-Chip Support for Up to 8 CPUs • Broadcast-Based 3-Hop I nvalidation Cache Coherence Protocol Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 36 AMD cHT Read Request Example Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 37 AMD cHT Read Request Example (cont.) Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 38 AMD cHT Read Request Example (cont.) Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 39 AMD cHT Read Request Example (cont.) Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 40 AMD cHT Read Request Example (cont.) Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 41 AMD cHT Read Request Example (cont.) Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 42 AMD cHT Read Request Example (cont.) Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 43 AMD cHT Read Request Example (cont.) Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 44 AMD cHT Read Request Example (cont.) Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 45 AMD cHT Read Request Example (cont.) Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 46 AMD cHT HT-Assist ( Probe Filter) • Old cHT Broadcast Protocol Broadcasts Probes to InvalidateCopies evenifMemoryLine isClean Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 47 AMD cHT HT-Assist (Probe Filter) (cont.) • Sparse Directory Cache Next to Memory Controller • Rule: I f a Line is Cached, it has an Entry in PF • Replacement Policy Makes Room for New Lines • Enhanced behavior: • No Probing for Uncached Lines • Directed Probe to Request Copy of Cached Line • Benefits: • Significantly Less Bandwidth Use • Shorter Access Latency, Mainly for Uncached Lines For Mor e Det ails: “Blade Computing with The AMD Magny-Cours Processor” Presented by - Pat Conway, Hot Chips 200 9 Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 48 AMD cHT HT-Assist (Probe Filter) (cont.) Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 49 Beyond Motherboards Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 50 HTX Connector • Low Latency CPU-to-High- Perf. Subsystem Direct Connect • Removes Performance Bottlenecks in Compute- Intensive Data Processing and Acceleration Functions • Com plem ents PCI - Class Interconnects • Link Splitting Capability Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 51 HTX Specification Evolution Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 52 New in HyperTransportTM 3 HT3 Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 53 HT3 - Link Splitting All Links or I ndividual Links 1x 4-Bit 2x 2-Bit 1x 8-Bit 2x 4-Bit 1x 16-Bit 2x 8-Bit 1x 32-Bit 2x 16-Bit Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 1 HT3 - Link Splitting (cont.) Extended SMP Topologies Enabled I/ O CPU CPU I/ O I/ O I/ O CPU CPU I/ O CPU CPU I/ O CPU CPU I/ O CPU CPU I/ O I/ O I/ O I/ O CPU CPU I/ O Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 55 HT3 – Dynamic Power Management • Dynamic Link Width, Clock Rate and Voltage Scaling • Partial Link Shutdow n via Link Splitting Hot Chips Symposium 2009 Copyright HyperTransport Consortium, 2009 56 HT3 – AC Mode (Optional – Enabled if Needed) • 8b/ 10b Encoding • Lower Bandwith, Higher Latency than DC Mode • DC/ AC Autoconfiguration • TX Equalization DC Mode AC Mode Decoupling Caps HT3 Spec HT3 Spec at < = 12 In at < = 3 Ft Tr a n sm i t w i t h Pr e - In-System Backplane and Post-Cursor De-Emphasis Chassis-
Recommended publications
  • ECESATUSB1 This Expresscard Power Esata Port Controller Card
    1 Port ExpressCard Power eSATA Controller Adapter Card StarTech ID: ECESATUSB1 This ExpressCard Power eSATA port controller card can be installed in an available ExpressCard 34/54 mm slot to provide a powered eSATA connection, and also alternatively provide either external SATA (data only) or USB 2.0 connectivity from one uniquely designed port if using with standard eSATA or USB devices. An ideal solution for using an eSATA SSD Flash drive on your laptop, the power eSATA card delivers both a high speed eSATA connection and power from the combined USB port. A versatile connectivity solution, the card features built-in port multiplier support, allowing multi-drive eSATA storage enclosures to be connected to the host computer using a single eSATA cable. Taking advantage of the transfer speed of eSATA connection and the 5V power output of the USB 2.0 port, the ExpressCard Power eSATA adapter is the perfect answer for connecting compatible mobile drive enclosures, similar to the built-in power eSATA port provided by the following laptop computers: Toshiba: Satellite E105, A350, Satellite Pro P300; Qosmio G50, X305, Portege A600, M750, R500, R600; and Tecra M10, R10, A10. Dell: Studio 15, 17; Latitude E6400, E6500; Precision M2400, M4400, M6400, M6400 Covet. Applications Connects to eSATA SSD Flash drives, such as OCZ Throttle, Kangaru e-Flash drives and Ridata Racer series flash drives Provides connectivity between Notebooks and PCs with ExpressCard slots to external drive enclosures with Power eSATA (eSATA+USB) port, or with regular eSATA
    [Show full text]
  • Computer Bus Characteristics
    Upendra Sharma (upsharma.in) Computer Bus A bus, in computing, is a set of physical connections (cables, printed circuits, etc.) which can be shared by multiple hardware components in order to communicate with one another. The purpose of buses is to reduce the number of "pathways" needed for communication between the components, by carrying out all communications over a single data channel. This is why the metaphor of a "data highway" is sometimes used. If only two hardware components communicate over the line, it is called a hardware port (such as a serial port or parallel port). Characteristics A bus is characterised by the amount of information that can be transmitted at once. This amount, expressed in bits, corresponds to the number of physical lines over which data is sent simultaneously. A 32-wire ribbon cable can transmit 32 bits in parallel. The term "width" is used to refer to the number of bits that a bus can transmit at once. Additionally, the bus speed is also defined by its frequency (expressed in Hertz), the number of data packets sent or received per second. Each time that data is sent or received is called a cycle. This way, it is possible to find the maximum transfer speed of the bus, the amount of data which it can transport per unit of time, by multiplying its width by its frequency. A bus with a width of 16 bits and a frequency of 133 MHz, therefore, has a transfer speed equal to: Upendra Sharma (upsharma.in) Types of Buses In reality, each bus is generally constituted of 50 to 100 distinct physical lines, divided into three subassemblies: The address bus (sometimes called the memory bus) transports memory addresses which the processor wants to access in order to read or write data.
    [Show full text]
  • 2 Port Flush Mount Expresscard 54Mm Superspeed USB 3.0 Card Adapter Startech ID: ECUSB3S254F
    2 Port Flush Mount ExpressCard 54mm SuperSpeed USB 3.0 Card Adapter StarTech ID: ECUSB3S254F The ECUSB3S254F 2-Port Flush Mount USB 3.0 ExpressCard Adapter uses a unique form factor design that allows it to sit fully in a standard 54mm laptop ExpressCard slot without sticking out. When inserted, the USB 3.0 ports provided by the ExpressCard adapter sit flush with the sides of the laptop, creating a seamless add-on that can be left installed even while on the move, without having to worry about impact damage to either the card or the ExpressCard slot. The SuperSpeed USB 3.0 Card supports data transfer rates of up to 5Gbps, and is backward compatible with USB 2.0 (at up to 480Mbps), or USB 1.1/1.0 at up to 12/1.5 Mbps respectively - the perfect laptop accessory for users to connect USB devices, both new and old. Applications Users who need USB connectivity, but do not need to swap between other ExpressCard adapters, so will leave the card installed in the card slot Mobile users who want to leave ExpressCard adapters installed, without worry about damaging the card or slot while on the move Connect high performance USB 3.0 external storage devices to a laptop Upgrade an older laptop with USB 3.0 connectivity Expand on your laptop expansion capabilities with additional USB ports Features Unique flush-mount form factor design Two SuperSpeed USB 3.0 compliant ports with support for transfer rates up to 5 Gbps Backward compatible with USB 2.0/1.x devices Compliant with USB 3.0 base specification 1.0 and xHCI specification 0.95 Compliant with
    [Show full text]
  • Hypertransport Extending Technology Leadership
    HyperTransport Extending Technology Leadership International HyperTransport Symposium 2009 February 11, 2009 Mario Cavalli General Manager HyperTransport Technology Consortium Copyright HyperTransport Consortium 2009 HyperTransport Extending Technology Leadership HyperTransport and Consortium Snapshot Industry Status and Trends HyperTransport Leadership Role February 11, 2009 Mario Cavalli General Manager HyperTransport Technology Consortium Copyright HyperTransport Consortium 2009 HyperTransport Snapshot Low Latency, High Bandwidth, High Efficiency Point-to-Point Interconnect Leadership CPU-to-I/O CPU-to-CPU CPU-to-Coprocessor Copyright HyperTransport Consortium 2009 Adopted by Industry Leaders in Widest Range of Applications than Any Other Interconnect Technology Copyright HyperTransport Consortium 2009 Snapshot Formed 2001 Controls, Licenses, Promotes HyperTransport as Royalty-Free Open Standard World Technology Leaders among Commercial and Academic Members Newly Elected President Mike Uhler VP Accelerated Computing Advanced Micro Devices Copyright HyperTransport Consortium 2009 Industry Status and Trends Copyright HyperTransport Consortium 2009 Global Economic Downturn Tough State of Affairs for All Industries Consumer Markets Crippled with Long-Term to Recovery Commercial Markets Strongly Impacted Copyright HyperTransport Consortium 2009 Consequent Business Focus Cost Effectiveness No Redundancy Frugality Copyright HyperTransport Consortium 2009 Downturn Breeds Opportunities Reinforced Need for More Optimized, Cost-Effective Computing
    [Show full text]
  • PDP-11 Bus Handbook (1979)
    The material in this document is for informational purposes only and is subject to change without notice. Digital Equipment Corpo­ ration assumes no liability or responsibility for any errors which appear in, this document or for any use made as a result thereof. By publication of this document, no licenses or other rights are granted by Digital Equipment Corporation by implication, estoppel or otherwise, under any patent, trademark or copyright. Copyright © 1979, Digital Equipment Corporation The following are trademarks of Digital Equipment Corporation: DIGITAL PDP UNIBUS DEC DECUS MASSBUS DECtape DDT FLIP CHIP DECdataway ii CONTENTS PART 1, UNIBUS SPECIFICATION INTRODUCTION ...................................... 1 Scope ............................................. 1 Content ............................................ 1 UNIBUS DESCRIPTION ................................................................ 1 Architecture ........................................ 2 Unibus Transmission Medium ........................ 2 Bus Terminator ..................................... 2 Bus Segment ....................................... 3 Bus Repeater ....................................... 3 Bus Master ........................................ 3 Bus Slave .......................................... 3 Bus Arbitrator ...................................... 3 Bus Request ....................................... 3 Bus Grant ......................................... 3 Processor .......................................... 4 Interrupt Fielding Processor .........................
    [Show full text]
  • A Unified Memory Network Architecture for In-Memory
    A Unified Memory Network Architecture for In-Memory Computing in Commodity Servers Jia Zhan∗, Itir Akgun∗, Jishen Zhao†, Al Davis‡, Paolo Faraboschi‡, Yuangang Wang§, and Yuan Xie∗ ∗University of California, Santa Barbara †University of California, Santa Cruz ‡HP Labs §Huawei Abstract—In-memory computing is emerging as a promising largely residing in main memory, they also pose significant paradigm in commodity servers to accelerate data-intensive challenges in DRAM scaling, especially for memory capacity. processing by striving to keep the entire dataset in DRAM. To If the large dataset cannot fit entirely in memory, it has to address the tremendous pressure on the main memory system, discrete memory modules can be networked together to form a be spilled to disk, thus causing sub-optimal performance due memory pool, enabled by recent trends towards richer memory to disk I/O contention. Unfortunately, with the current DDRx interfaces (e.g. Hybrid Memory Cubes, or HMCs). Such an inter- based memory architecture, the scaling of DRAM capacity memory network provides a scalable fabric to expand memory falls far behind the growth of dataset size of in-memory capacity, but still suffers from long multi-hop latency, limited computing applications [12]. A straightforward approach to bandwidth, and high power consumption—problems that will continue to exacerbate as the gap between interconnect and increase memory capacity is to add more CPU sockets for transistor performance grows. Moreover, inside each memory additional memory channels. However, this will introduce module, an intra-memory network (NoC) is typically employed to significant hardware cost [13]. Moreover, in a multi-socket connect different memory partitions.
    [Show full text]
  • Model 225E MIDI Decoder / Preset Manager the 225E MIDI Decoder Translates Incoming MIDI Messages Into Analog Voltages and Pulses
    Description - model 225e MIDI Decoder / Preset Manager The 225e MIDI decoder translates incoming MIDI messages into analog voltages and pulses. These control signals appear on three kinds of "busses" for subsequent distribution to 200e series modules. Four busses are routed internally to 200e series modules that respond directly to MIDI messages. Four more handle only note messages, decoding them into front panel voltages that represent pitches, velocities and pulses with sustain information. An additional six busses handle controller and clock messages (two controllers each), which are translated into standard 0-10 volt control voltages. Busses: Note and Controller message assigments to Pulses and CVs. Note messages are applied to busses A,B,C,D,E,F,G, and H; Continuous Controller and Clock messages are handled by busses J, K, L, M, N, and P. Each bus responds to one selected MIDI channel. But more than one bus may be assigned to the same channel, allowing multiple controllers CVs and note messages to be derived from a single MIDI channel. First select a bus for assignment use "bus select" to illuminate the LED next to the desired bus. MIDI NoteOn and pitch translation: Busses A-H There are two sets of Note messages busses: A-D and E-H. Internal busses will be explained in more detail later. For these busses the following display will appear: Bus:A Ch:15 Poly Tr:24 Ft:75 Vl:/ With the cursor buttons, select a data type. Use the knob to the right of the cursor buttons to enter data for the following fields: Ch: MIDI Channel.
    [Show full text]
  • Introduction to AGP-8X
    W H I T E P A P E R Introduction to AGP-8X Allan Chen Jessie Johnson Angel Suhrstedt ADVANCED MICRO DEVICES, INC. One AMD Place Sunnyvale, CA 94088 Page 1 Introduction to AGP-8X May 10, 2002 W H I T E P A P E R Background Tremendous efforts have been made to improve computing system performance through increases in CPU processing power; however, I/O bottlenecks throughout the computing platform can limit the system performance. To eliminate system bottlenecks, AMD has been working with industry leaders to implement innovative technologies including AGP, DDR SDRAM, USB2.0, and HyperTransport™ technology. This white paper will provide an overview of the latest graphics subsystem innovation: AGP-8X. Introduction The Accelerated Graphics Port (AGP) was developed as a high-performance graphics interface. It alleviates the PCI graphics bottleneck by providing high bandwidth throughput and direct access to system memory. The new AGP3.0 specification adds an 8X mode definition, which doubles the maximum data transfer rate from the previous high reached with AGP-4X by doubling the amount of data transferred per AGP bus cycle. Figure 1 shows the graphic interface bandwidth performance evolution from PCI to AGP-8X. In this figure, AGP3.0 refers to the specific AGP interface specification. AGP-1X, AGP-2X, AGP-4X and AGP-8X represent the data transfer speed mode. Graphics Interface Peak Bandwidth 2500 2000 1500 1000 Bandwidth (MB/s) 500 0 PCI AGP1X AGP2X AGP4X AGP8X AGP 1.0 Specification AGP 2.0 Specification AGP 3.0 Specification Figure 1: Available bandwidth of different graphic interfaces.
    [Show full text]
  • Computing Platforms Chapter 4
    Computing Platforms Chapter 4 COE 306: Introduction to Embedded Systems Dr. Abdulaziz Tabbakh Computer Engineering Department College of Computer Sciences and Engineering King Fahd University of Petroleum and Minerals [Adapted from slides of Dr. A. El-Maleh, COE 306, KFUPM] Next . Basic Computing Platforms The CPU bus Direct Memory Access (DMA) System Bus Configurations ARM Bus: AMBA 2.0 Memory Components Embedded Platforms Platform-Level Performance Computing Platforms COE 306– Introduction to Embedded System– KFUPM slide 2 Embedded Systems Overview Actuator Output Analog/Digital Sensor Input Analog/Digital CPU Memory Embedded Computer Computing Platforms COE 306– Introduction to Embedded System– KFUPM slide 3 Computing Platforms Computing platforms are created using microprocessors, I/O devices, and memory components A CPU bus is required to connect the CPU to other devices Software is required to implement an application Embedded system software is closely tied to the hardware Computing Platform: hardware and software Computing Platforms COE 306– Introduction to Embedded System– KFUPM slide 4 Computing Platform A typical computing platform includes several major hardware components: The CPU provides basic computational facilities. RAM is used for program and data storage. ROM holds the boot program and some permanent data. A DMA controller provides direct memory access capabilities. Timers are used by the operating system A high-speed bus, connected to the CPU bus through a bridge, allows fast devices to communicate efficiently with the rest of the system. A low-speed bus provides an inexpensive way to connect simpler devices and may be necessary for backward compatibility as well. Computing Platforms COE 306– Introduction to Embedded System– KFUPM slide 5 Platform Hardware Components Computer systems may have one or more bus Buses are classified by their overall performance: lows peed, high- speed.
    [Show full text]
  • AMD Opteron™ 4000 Series Platform Quick Reference Guide
    AMD Opteron™ 4000 Series Platform Quick Reference Guide The AMD Opteron™ 4100 Series processor, the world’s lowest power per core processor1, sets the foundation for cloud workloads and affordability for mainstream infrastructure servers with prices starting at $992. AMD Opteron™ 4100 Series Processor END USER BENEFITS Outstanding Performance-Per-Watt > Designed from the ground up to handle demanding server workloads at the lowest available energy draw, beating the competition by as much as 40% (per/core).1 Business Value > The world’s first 1P and 2P capable processor at sub $100 pricing.2 Easy to Purchase and Operate > Scalable solutions with feature, component and platform consistency. PRODUCT FEATURES New AMD-P 2.0 Power Savings Features: > Ultra-low power platforms provide power efficiency beyond just the processor, 3 SOUNDBITE for both 1P and 2P server configurations. THE WORLD’S LOWEST POWER PER CORE SERVER PROCESSOR1 > APML (Advanced Platform Management Link)4 provides an interface for processor and systems management monitoring and controlling of system resources such as platform power Quick Features consumption via p-state limits and CPU thermals to closely monitor power and cooling. AMD-P 2.0: > Link Width PowerCap which changes all 16-bit links to 8-bit links5 can help power conscious Ultra-low power platform > customers improve performance-per-watt. > Advanced Platform Management Link (APML)4 > AMD CoolSpeed Technology reduces p-states when a temperature limit is reached to allow a > Link Width PowerCap server to operate if the processor’s thermal environment exceeds safe operational limits. > AMD CoolSpeed Technology > When C1E5,6 is enabled, the cores, southbridge and memory controller enter a sleep state that > C1E6 can equate to significant power savings in the datacenter depending on system configuration.
    [Show full text]
  • An FPGA Based Verification Platform for Hypertransport 3.X
    An FPGA based Verification Platform for HyperTransport 3.x Heiner Litz Holger Froening Maximilian Thuermer Ulrich Bruening University of Heidelberg Computer Architecture Group Germany {heiner.litz, holger.froening, ulrich.bruening}@ziti.uni-heidelberg.de ABSTRACT ware (HW/SW) co-verification Field Programmable Gate In this paper we present a verification platform designed Arrays (FPGAs) have several advantages compared to for HyperTransport 3.x (HT3) applications. It is intended to Electronic Design Automation (EDA) tools. FPGAs allow be used in computing environments in which it is directly to run prototypes in real world environments, in which both connected over a HyperTransport link to the main CPUs. hardware modules and software layers can be verified in an No protocol conversions or intermediate bridges are neces- early prototyping stage. However, the drawback is that sary, which results in a very low latency. An FPGA tech- FPGA-based verification platforms have to be specially nology is chosen due to of its reconfigurability. The main designed for certain applications, in particular if the appli- challenge of this work is the implementation of an HT3 cation is based on a new technology. link using FPGA technology. We provide a design space The verification platform presented here connects to a exploration in order to find the most suitable I/O technol- HyperTransport (HT) link. HyperTransport 2.x (HT2) ogy. This includes verification of the HyperTransport-spe- [1][2] is an established technology for chip-to-chip connec- cific signal levelling, the development of an I/O tions offering a very low latency and a high bandwidth. It is architecture supporting a link width of up to 16 lanes with used in all major AMD processors, including the Opteron source synchronous clocking and a clocking scheme to [3].
    [Show full text]
  • MPC107 PCI Bridge/Integrated Memory Controller Fact Sheet
    Freescale Semiconductor, Inc. Fact Sheet MPC107 PCI BRIDGE MEMORY CONTROLLER The MPC107 PCI Bridge/Integrated Memory Controller provides a bridge between the INTEGRATED MEMORY CONTROLLER Peripheral Component Interconnect (PCI) bus and Motorola MPC6xx, MPC7xx, and The memory interface controls processor and PCI . MPC74xx microprocessors. PCI support allows system designers to design systems interactions to main memory. It supports a variety of . quickly using peripherals already designed for PCI and the other standard interfaces programmable timing supporting DRAM (FPM, EDO), c available in the personal computer hardware environment. The MPC107 provides many SDRAM, and ROM/Flash ROM configurations, up to n of the other necessities for embedded applications, including a high-performance speeds of 100 MHz. I , memory controller and dual-processor support; two-channel flexible DMA controller; an r interrupt controller; an I2O-ready message unit; an inter-integrated circuit controller PCI BUS SUPPORT o (I2C); and low-skew clock drivers. The MPC107 contains an Embedded Programmable The MPC107 PCI interface is designed to connect the t Interrupt Controller (EPIC) featuring five hardware interrupts (IRQs) as well as 16 serial c processor and memory buses to the PCI local bus without u interrupts along with four timers. The MPC107 uses an advanced, 2.5V HiP3 process the need for “glue” logic at speeds up to 66 MHz. The d technology and is fully compatible with TTL devices. MPC107 acts as either a master or slave device on the n o MOTOROLA MPC107 BLOCK DIAGRAM c i m 60x Bus e S Memory Data Data Path e l ECC/Parity a Unit PIU c I O s 2 Memory e Add/Cntl e CCU MCU r DMA F 2 Mem Clocks I C DLL I2C PCIU IRQs Clock In EPIC PLL PCI Clocks ATU Arbiter FO Buffer Timers PCI Bus REQ/GNT OSC For More Information On This Product, Go to: www.freescale.com Freescale Semiconductor, Inc.
    [Show full text]