Paving the Path for Heterogeneous Memory Adoption in Production Systems

Total Page:16

File Type:pdf, Size:1020Kb

Paving the Path for Heterogeneous Memory Adoption in Production Systems Paving the Path for Heterogeneous Memory Adoption in Production Systems by Neha Agarwal A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy (Computer Science and Engineering) in The University of Michigan 2017 Doctoral Committee: Associate Professor Thomas F. Wenisch, Chair Professor Ella M. Atkins Professor Peter M. Chen Assistant Professor Ronald G. Dreslinski Jr. Neha Agarwal [email protected] ORCID iD: 0000-0001-9833-6538 c Neha Agarwal 2017 2/10/2017 thesis-tag-cloudsans-serif – Word cloud – WordItOut [/] Create [/word­cloud/create] thesis­tag­cloudsans­serif Thermostat heterogeneous information selective interconnect technologies placement hardware overhead DDR use CPU rate page systems application coherence using high ratio policy GDDR two memories pages without while shown capacity each support virtual TLB threshold OS client only between across GPU hot any run cold GPUs per shared Online slow some Linux software cloud memory caching huge KB request number access caches Section table cost first CPU–GPU accessed higher fault range workloads runtime applications because cache data coalescing address requests latency more shows slowdown footprint Computer accesses mechanism throughput total Conference fraction performance DRAM degradation Architecture expansion system bandwidth BW­AWARE migration Like 0 Tweet Share [https://www.addtoany.com/share#url=https%3A%2F%2Fworditout.com%2Fword­cloud%2F2036243%2Fprivate%2F92b78a883823cd01e3b1de52494d279a&title=thesis­tag­cloudsans­serif%20%E2%80%93%20Word%20cloud%20%E2%80%93%20WordItOut] Download Fullscreen A great gift idea You may now get this word cloud on many items, such as T­shirts, mugs, cards, bags and even more! They make great custom gifts for Thermostat heterogeneous information selective interconnect technologies placement hardware overhead DDR use CPU rate page systems application coherence using high [/word­ GDDR ratio policy two memories pages without while someone special as well as personalised shown capacity each support virtual TLB threshold OS client only between across GPU hot any run cold GPUs per shared Online slow some Linux software cloud memory caching huge KB request number access caches cloud/2036243/private/92b78a883823cd01e3b1de52494d279a/goto/zazzle.com]Section table cost first accessed CPU–GPU higher fault range workloads runtime applications because cache data coalescing address requests latency more shows slowdown footprint Computer presents for yourself. accesses mechanism throughput total Conference fraction performance DRAM degradation Architecture expansion system bandwidth BW­AWARE migration Find out more [/word­ cloud/2036243/private/92b78a883823cd01e3b1de52494d279a/goto/zazzle.com] If you do buy something with these affiliate links, you help WordItOut continue to run for free at no extra cost to you. Area: United States In association with Zazzle.com Embed To embed this word cloud on your own website, just copy and paste the following code: <!­­ Creative Commons Licence (by­nc­ nd). See worditout.com for details ­­> <div style='width:auto;height:auto;'> <!­­ You may use this wrapping div to [https://creativecommons.org/licenses/by­nc­ nd/2.0/uk/deed.en_GB] Word clouds are licensed under a Creative Commons Licence https://worditout.com/word-cloud/2036243/private/92b78a883823cd01e3b1de52494d279a 1/2 To – Family ii ACKNOWLEDGEMENTS First and foremost I want to sincerely thank Tom, my PhD advisor, who gave me the opportunity to pursue the doctoral program with his guidance. Tom is a fantastic advisor, a statement that I have been saying time and again especially new students seeking an advisor. In Indian culture we have a special place and respect for the “guru” and I feel that I was fortunate enough to find Tom and to be able to associate with him as my “guru”. Tom is a highly energetic, optimistic, professional, and extremely knowledgeable person. His advising style is very adaptive and he let’s each student learn and work at his/her own pace. He does not create work pressure, but if you are going off tracks he will for sure warn you. He is an extremely adaptive and flexible advisor be it technically, administratively, or logistically. Because of his support I was able to explore different positions in industry and understand the contrasting but complementary research approaches in academia versus industry. With his guidance I have learnt to deal with rejections and failures with a positive attitude, which is an instrumental medicine for PhD journey. No matter how many rejections you get, his philosophy of keep trying to improve and doing excellent quality research is what I found most useful as his student. I am thankful to University of Michigan, CSE department to give me a wonderful opportunity of associating with Tom, as my advisor. I also want to take this chance to thank all of my collaborators for their contributions in making my ideas more concrete and provide me the perspective that I lacked before. I especially want to mention about the guidance and support I got from David Nellans. Dave was my internship mentor at NVIDIA. His approach of conducting product influential research helped me understand impact of research on products and business utility for a company. Dave is also a very good friend who boosts my moral and motivates me to keep doing good work. Andres´ Lagar-Cavilla is another important name in my technical career, because of whom I got interested in Linux and operating systems. His commitment to help and support his peers inspires me. His help during my internship at Google was the stepping stone for my success in my internship project. I think coming into the PhD program was one of the best decisions of my life. I got many opportunities to travel in the last six years, exploring different countries, cultures, iii and lifestyles. Travel gave me a perspective to meaning of life. I could go out of my comfort zone, experience people’s generosity, overcome difficulty of not knowing whereabouts of a place. I realized that world is a much bigger and beautiful place and that experience matters more than success. Travel inspired me to explore new avenues in my research, become comfortable at switching projects and keep learning new things. My parents – Madhu Agarwal and Bal Krishna Agarwal – are the reason whatever I am today. Their constant guidance and support in life has been instrumental in my writing this acknowledgement section today. They have provided me the shade and comfort at each and every single step in my life even when I didn’t ask for it. They always understand what I need and are always there for me. The feeling of having them no matter what hap- pened in my professional or personal life is the biggest gift I have gotten in my life. When things were not going the way I wanted them to go for very long time and I was doubting myself all the time, mummy papa were the one who always filled me with confidence and encouraged me to try even harder. I have called them several times late at nights to talk about how worried I am, how things are not happening even though I am doing all I can, they are always there to listen to me quietly and help me move forward. I can’t imagine doing anything without their support. Thanks is a small word for what my parents have done for me. I can just say that I am lucky enough to have them in my life. Mummy–Papa this dissertation is your hard work and you confidence in me to be able to sincerely work and contribute to the world. Next, I want to talk about my brother, Shyam Mohan Agarwal. He is the person who loves me so much that at times I think what did I do to receive that much love. He is the one because of whom I am able to fulfill my dream of studying in the best institutes of knowledge. His encouragement and belief in me is why I have the courage to try out new things fearlessly. He looked something in me since childhood and he knew that I would be a scholar. It has been like he has foreseen what I will become and then gave me a push to go and get it. With every achievement of mine it feels like he actually won something. He is the one who enjoys my success the most. Without him experiencing the joy of what I do my work is incomplete. He is a true inspiration to me, without whom I am not sure how I would have a direction to pursue something that I love most doing. I also want to thank my sister-in-law, Khyati Agarwal, who has been so nice to accommodate all of us in her life and continued to make our family loving and happy. I have also been very fortunate to have come across several fantastic people and be friends with them. Gaurav, Pooja, Kataria, Gauri, Tanvi, Mohit, Kunal, Pulkit are some names very close to my heart that I know will come for me if I need them at any point in life. They have challenged me in several ways, making me a better person with every iv interaction. Countless nights we have just sat together and just discussed some weird mystery of life or just cribbed about how things are unfair. These people are treasures I have found in my life and have made my life so comfortable and loving. They have played a significant role by keeping me sane during the most self-doubting periods of PhD journey. All of my seniors Gaurav Chadha, Ankit Sethia, Daya Khudia, and Abhayendra Singh are some of the people who helped me acclimatize in Ann Arbor and the department during my foundational years of my PhD.
Recommended publications
  • Supermicro GPU Solutions Optimized for NVIDIA Nvlink
    SuperServers Optimized For NVIDIA® Tesla® GPUs Maximizing Throughput and Scalability 16 Tesla® V100 GPUs With NVLink™ and NVSwitch™ Most Powerful Solution for Deep Learning Training • New Supermicro NVIDIA® HGX-2 based platform • 16 Tesla® V100 SXM3 GPUs (512GB total GPU memory) • 16 NICs for GPUDirect RDMA • 16 hot-swap NVMe drive bays • Fully configurable to order SYS-9029GP-TNVRT 8 Tesla® V100 GPUs With NVLink™ 4 Tesla® V100 GPUs With NVLink™ SYS-1029GQ-TVRT SYS-4029GP-TVRT www.supermicro.com/GPU March 2019 Maximum Acceleration for AI/DL Training Workloads PERFORMANCE: Highest Parallel peak performance with NVIDIA Tesla V100 GPUs THROUGHPUT: Best in class GPU-to-GPU bandwidth with a maximum speed of 300GB/s SCALABILITY: Designed for direct interconections between multiple GPU nodes FLEXIBILITY: PCI-E 3.0 x16 for low latency I/O expansion capacity & GPU Direct RDMA support DESIGN: Optimized GPU cooling for highest sustained parallel computing performance EFFICIENCY: Redundant Titanium Level power supplies & intelligent cooling control Model SYS-1029GQ-TVRT SYS-4029GP-TVRT • Dual Intel® Xeon® Scalable processors with 3 UPI up to • Dual Intel® Xeon® Scalable processors with 3 UPI up to 10.4GT/s CPU Support 10.4GT/s • Supports up to 205W TDP CPU • Supports up to 205W TDP CPU • 8 NVIDIA® Tesla® V100 GPUs • 4 NVIDIA Tesla V100 GPUs • NVIDIA® NVLink™ GPU Interconnect up to 300GB/s GPU Support • NVIDIA® NVLink™ GPU Interconnect up to 300GB/s • Optimized for GPUDirect RDMA • Optimized for GPUDirect RDMA • Independent CPU and GPU thermal zones
    [Show full text]
  • High Performance Computing and AI Solutions Portfolio
    Brochure High Performance Computing and AI Solutions Portfolio Technology and expertise to help you accelerate discovery and innovation Discovery and innovation have always started with great minds Go ahead. dreaming big. As artificial intelligence (AI), high performance computing (HPC) and data analytics continue to converge and Dream big. evolve, they are fueling the next industrial revolution and the next quantum leap in human progress. And with the help of increasingly powerful technology, you can dream even bigger. Dell Technologies will be there every step of the way with the technology you need to power tomorrow’s discoveries and the expertise to bring it all together, today. 463 exabytes The convergence of HPC and AI is driven by data. The data‑driven age is dramatically reshaping industries and reinventing the future. As of data will be created each day by 20251 vast amounts of data pour in from increasingly diverse sources, leveraging that data is both critical and transformational. Whether you’re working to save lives, understand the universe, build better machines, neutralize financial risks or anticipate customer sentiment, 44X ROI data informs and drives decisions that impact the success of your organization — and Average return on investment (ROI) shapes the future of our world. for HPC2 AI, HPC and data analytics are technologies designed to unlock the value of your data. While they have long been treated as separate, the three technologies are converging as 83% of CIOs it becomes clear that analytics and AI are both big‑data problems that require the power, Say they are investing in AI scalable compute, networking and storage provided by HPC.
    [Show full text]
  • BRKIOT-2394.Pdf
    Unlocking the Mystery of Machine Learning and Big Data Analytics Robert Barton Jerome Henry Distinguished Architect Principal Engineer @MrRobbarto Office of the CTAO CCIE #6660 @wirelessccie CCDE #2013::6 CCIE #24750 CWNE #45 BRKIOT-2394 Cisco Webex Teams Questions? Use Cisco Webex Teams to chat with the speaker after the session How 1 Find this session in the Cisco Events Mobile App 2 Click “Join the Discussion” 3 Install Webex Teams or go directly to the team space 4 Enter messages/questions in the team space BRKIOT-2394 © 2020 Cisco and/or its affiliates. All rights reserved. Cisco Public 3 Tuesday, Jan. 28th Monday, Jan. 27th Wednesday, Jan. 29th BRKIOT-2600 BRKIOT-2213 16:45 Enabling OT-IT collaboration by 17:00 From Zero to IOx Hero transforming traditional industrial TECIOT-2400 networks to modern IoT Architectures IoT Fundamentals 08:45 BRKIOT-1618 Bootcamp 14:45 Industrial IoT Network Management PSOIOT-1156 16:00 using Cisco Industrial Network Director Securing Industrial – A Deep Dive. Networks: Introduction to Cisco Cyber Vision PSOIOT-2155 Enhancing the Commuter 13:30 BRKIOT-1775 Experience - Service Wireless technologies and 14:30 BRKIOT-2698 BRKIOT-1520 Provider WiFi at the Use Cases in Industrial IOT Industrial IoT Routing – Connectivity 12:15 Cisco Remote & Mobile Asset speed of Trains and Beyond Solutions PSOIOT-2197 Cisco Innovates Autonomous 14:00 TECIOT-2000 Vehicles & Roadways w/ IoT BRKIOT-2497 BRKIOT-2900 Understanding Cisco's 14:30 IoT Solutions for Smart Cities and 11:00 Automating the Network of Internet Of Things (IOT) BRKIOT-2108 Communities Industrial Automation Solutions Connected Factory Architecture Theory and 11:00 Practice PSOIOT-2100 BRKIOT-1291 Unlock New Market 16:15 Opening Keynote 09:00 08:30 Opportunities with LoRaWAN for IOT Enterprises Embedded Cisco services Technologies IOT IOT IOT Track #CLEMEA www.ciscolive.com/emea/learn/technology-tracks.html Cisco Live Thursday, Jan.
    [Show full text]
  • Hypertransport Extending Technology Leadership
    HyperTransport Extending Technology Leadership International HyperTransport Symposium 2009 February 11, 2009 Mario Cavalli General Manager HyperTransport Technology Consortium Copyright HyperTransport Consortium 2009 HyperTransport Extending Technology Leadership HyperTransport and Consortium Snapshot Industry Status and Trends HyperTransport Leadership Role February 11, 2009 Mario Cavalli General Manager HyperTransport Technology Consortium Copyright HyperTransport Consortium 2009 HyperTransport Snapshot Low Latency, High Bandwidth, High Efficiency Point-to-Point Interconnect Leadership CPU-to-I/O CPU-to-CPU CPU-to-Coprocessor Copyright HyperTransport Consortium 2009 Adopted by Industry Leaders in Widest Range of Applications than Any Other Interconnect Technology Copyright HyperTransport Consortium 2009 Snapshot Formed 2001 Controls, Licenses, Promotes HyperTransport as Royalty-Free Open Standard World Technology Leaders among Commercial and Academic Members Newly Elected President Mike Uhler VP Accelerated Computing Advanced Micro Devices Copyright HyperTransport Consortium 2009 Industry Status and Trends Copyright HyperTransport Consortium 2009 Global Economic Downturn Tough State of Affairs for All Industries Consumer Markets Crippled with Long-Term to Recovery Commercial Markets Strongly Impacted Copyright HyperTransport Consortium 2009 Consequent Business Focus Cost Effectiveness No Redundancy Frugality Copyright HyperTransport Consortium 2009 Downturn Breeds Opportunities Reinforced Need for More Optimized, Cost-Effective Computing
    [Show full text]
  • NVIDIA Gpudirect RDMA (GDR) Host Memory Host Memory • Pipeline Through Host for Large Msg 2 IB IB 4 2
    Efficient and Scalable Communication Middleware for Emerging Dense-GPU Clusters Ching-Hsiang Chu Advisor: Dhabaleswar K. Panda Network Based Computing Lab Department of Computer Science and Engineering The Ohio State University, Columbus, OH Outline • Introduction • Problem Statement • Detailed Description and Results • Broader Impact on the HPC Community • Expected Contributions Network Based Computing Laboratory SC 19 Doctoral Showcase 2 Trends in Modern HPC Architecture: Heterogeneous Multi/ Many-core High Performance Interconnects Accelerators / Coprocessors SSD, NVMe-SSD, Processors InfiniBand, Omni-Path, EFA high compute density, NVRAM <1usec latency, 100Gbps+ Bandwidth high performance/watt Node local storage • Multi-core/many-core technologies • High Performance Storage and Compute devices • High Performance Interconnects • Variety of programming models (MPI, PGAS, MPI+X) #1 Summit #2 Sierra (17,280 GPUs) #8 ABCI #22 DGX SuperPOD (27,648 GPUs) #10 Lassen (2,664 GPUs) (4,352 GPUs) (1,536 GPUs) Network Based Computing Laboratory SC 19 Doctoral Showcase 3 Trends in Modern Large-scale Dense-GPU Systems • Scale-up (up to 150 GB/s) • Scale-out (up to 25 GB/s) – PCIe, NVLink/NVSwitch – InfiniBand, Omni-path, Ethernet – Infinity Fabric, Gen-Z, CXL – Cray Slingshot Network Based Computing Laboratory SC 19 Doctoral Showcase 4 GPU-enabled HPC Applications Lattice Quantum Chromodynamics Weather Simulation Wave propagation simulation Fuhrer O, Osuna C, Lapillonne X, Gysi T, Bianco M, Schulthess T. Towards GPU-accelerated operational weather
    [Show full text]
  • Dell EMC Poweredge C4140 Technical Guide
    Dell EMC PowerEdge C4140 Technical Guide Regulatory Model: E53S Series Regulatory Type: E53S001 Notes, cautions, and warnings NOTE: A NOTE indicates important information that helps you make better use of your product. CAUTION: A CAUTION indicates either potential damage to hardware or loss of data and tells you how to avoid the problem. WARNING: A WARNING indicates a potential for property damage, personal injury, or death. © 2017 - 2019 Dell Inc. or its subsidiaries. All rights reserved. Dell, EMC, and other trademarks are trademarks of Dell Inc. or its subsidiaries. Other trademarks may be trademarks of their respective owners. 2019 - 09 Rev. A00 Contents 1 System overview ......................................................................................................................... 5 Introduction............................................................................................................................................................................ 5 New technologies.................................................................................................................................................................. 5 2 System features...........................................................................................................................7 Specifications......................................................................................................................................................................... 7 Product comparison.............................................................................................................................................................
    [Show full text]
  • HPE Apollo 6500 Gen10 System Overview
    QuickSpecs HPE Apollo 6500 Gen10 System Overview HPE Apollo 6500 Gen10 System The ability of computers to autonomously learn, predict, and adapt using massive datasets is driving innovation and competitive advantage across many industries and applications. The HPE Apollo 6500 Gen10 System is an ideal HPC and Deep Learning platform providing unprecedented performance with industry leading GPUs, fast GPU interconnect, high bandwidth fabric and a configurable GPU topology to match your workloads. The system with rock-solid RAS features (reliable, available, secure) includes up to eight high power GPUs per server tray (node), NVLink 2.0 for fast GPU-to-GPU communication, Intel® Xeon® Scalable Processors support, choice of up to four high-speed / low latency fabric adapters, and the ability to optimize your configurations to match your workload and choice of GPU. And while the HPE Apollo 6500 Gen10 System is ideal for deep learning workloads, the system is suitable for complex high performance computing workloads such as simulation and modeling. Eight GPU per server for faster and more economical deep learning system training compared to more servers with fewer GPU each. Keep your researchers productive as they iterate on a model more rapidly for a better solution, in less time. Now available with NVLink 2.0 to connect GPUs at up to 300 GB/s for the world’s most powerful computing servers. HPC and AI models that would consume days or weeks can now be trained in a few hours or minutes. Front View –with two HPE DL38X Gen10 Premium 6 SFF SAS/SATA + 2 NVMe drive cages shown 1.
    [Show full text]
  • A Unified Memory Network Architecture for In-Memory
    A Unified Memory Network Architecture for In-Memory Computing in Commodity Servers Jia Zhan∗, Itir Akgun∗, Jishen Zhao†, Al Davis‡, Paolo Faraboschi‡, Yuangang Wang§, and Yuan Xie∗ ∗University of California, Santa Barbara †University of California, Santa Cruz ‡HP Labs §Huawei Abstract—In-memory computing is emerging as a promising largely residing in main memory, they also pose significant paradigm in commodity servers to accelerate data-intensive challenges in DRAM scaling, especially for memory capacity. processing by striving to keep the entire dataset in DRAM. To If the large dataset cannot fit entirely in memory, it has to address the tremendous pressure on the main memory system, discrete memory modules can be networked together to form a be spilled to disk, thus causing sub-optimal performance due memory pool, enabled by recent trends towards richer memory to disk I/O contention. Unfortunately, with the current DDRx interfaces (e.g. Hybrid Memory Cubes, or HMCs). Such an inter- based memory architecture, the scaling of DRAM capacity memory network provides a scalable fabric to expand memory falls far behind the growth of dataset size of in-memory capacity, but still suffers from long multi-hop latency, limited computing applications [12]. A straightforward approach to bandwidth, and high power consumption—problems that will continue to exacerbate as the gap between interconnect and increase memory capacity is to add more CPU sockets for transistor performance grows. Moreover, inside each memory additional memory channels. However, this will introduce module, an intra-memory network (NoC) is typically employed to significant hardware cost [13]. Moreover, in a multi-socket connect different memory partitions.
    [Show full text]
  • The Computer That Could Be Smarter Than Us Cognitive Computing
    outthink limits The Computer that could be smarter than us Cognitive Computing Ingolf Wittmann Technical Director Leader of HPC Kinetica - Unstructured Databases What is Kinetica? Unparalleled acceleration Kinetica’s in-memory database powered by graphics processing units (GPUs) was built from the ground up to deliver truly real-time insights on data in 2.5x motion: orders of magnitude faster Bandwidth performance (potentially 100, 1000X) at 100100 ticktick QueryQuery Time:Time: CompetingCompeting SystemSystem PCI-EPCI-E x16x16 3.03.0 10% to 25% of the cost of traditional An industry first — POWER8 with data platforms. NVIDIA NVLink delivers 2.5X the Calculation* bandwidth to GPU accelerators, Data Transfer allowing you to experience Kinetica at 73 ticks 27 ticks the speed it was intended compared to x86 based systems. What are the Key Markets? 65% Reduction • Retail: Inventory Mgt, BI, Apps, Big Data tools, HPA Real-time results • Distribution / Logistics : Supply Chain Data Transfer Calculation* Mgt 26 ticks 14 ticks • Financial Services : Fraud Detection, AML 4040 ticktick QueryQuery Time:Time: S822LCS822LC forfor HPC,HPC, NVLinkNVLink 10x • Ad-Tech : More Targeted Marketing * Includes non-overlapping: CPU, GPU, and idle times. Performance • IoT : End Point Management, RFID With the unique capabilities of Tesla 65% reduction in data transfer time (3X improvement) in for P100 + POWER8, Kinetica has 2.4x POWER8 with NVLink the performance of competing Kinetica GPU-accelerated DB PCIe x16 3.0/x86 System System systems enabling you to analyze and • Less data-induced latency in all applications Xeon E5-2640 v4 Power Systems S822LC visualize large datasets in • with 4 Tesla K80s : Unique to POWER8 with NVLink with 4 Tesla P100s: milliseconds vs.
    [Show full text]
  • NVIDIA Geforce RTX 2080 User Guide | 3 Introduction
    2080 TABLE OF CONTENTS 01 INTRODUCTION ..................................................................................... 3 About This Guide ................................................................................................................................ 3 Minimum System Requirements ....................................................................................................... 4 02 UNPACKING .......................................................................................... 5 Equipment .......................................................................................................................................... 6 03 Hardware Installation ............................................................................. 7 Safety Instructions ............................................................................................................................. 7 Before You Begin ................................................................................................................................ 8 Installing the GeForce Graphics Card ............................................................................................... 8 04 SOFTWARE INSTALLATION ................................................................... 12 GeForce Experience Software Installation ...................................................................................... 12 05 CONFIGURING NVLink ......................................................................... 14 06 CONFIGURING HDMI ...........................................................................
    [Show full text]
  • Introduction to AGP-8X
    W H I T E P A P E R Introduction to AGP-8X Allan Chen Jessie Johnson Angel Suhrstedt ADVANCED MICRO DEVICES, INC. One AMD Place Sunnyvale, CA 94088 Page 1 Introduction to AGP-8X May 10, 2002 W H I T E P A P E R Background Tremendous efforts have been made to improve computing system performance through increases in CPU processing power; however, I/O bottlenecks throughout the computing platform can limit the system performance. To eliminate system bottlenecks, AMD has been working with industry leaders to implement innovative technologies including AGP, DDR SDRAM, USB2.0, and HyperTransport™ technology. This white paper will provide an overview of the latest graphics subsystem innovation: AGP-8X. Introduction The Accelerated Graphics Port (AGP) was developed as a high-performance graphics interface. It alleviates the PCI graphics bottleneck by providing high bandwidth throughput and direct access to system memory. The new AGP3.0 specification adds an 8X mode definition, which doubles the maximum data transfer rate from the previous high reached with AGP-4X by doubling the amount of data transferred per AGP bus cycle. Figure 1 shows the graphic interface bandwidth performance evolution from PCI to AGP-8X. In this figure, AGP3.0 refers to the specific AGP interface specification. AGP-1X, AGP-2X, AGP-4X and AGP-8X represent the data transfer speed mode. Graphics Interface Peak Bandwidth 2500 2000 1500 1000 Bandwidth (MB/s) 500 0 PCI AGP1X AGP2X AGP4X AGP8X AGP 1.0 Specification AGP 2.0 Specification AGP 3.0 Specification Figure 1: Available bandwidth of different graphic interfaces.
    [Show full text]
  • NVIDIA Geforce RTX 2080 User Guide | 3 Introduction
    2080 TABLE OF CONTENTS 01 INTRODUCTION ..................................................................................... 3 About This Guide ................................................................................................................................ 3 Minimum System Requirements ....................................................................................................... 4 02 UNPACKING .......................................................................................... 5 Equipment .......................................................................................................................................... 6 03 Hardware Installation ............................................................................. 7 Safety Instructions ............................................................................................................................. 7 Before You Begin ................................................................................................................................ 8 Installing the GeForce Graphics Card ............................................................................................... 8 04 SOFTWARE INSTALLATION ................................................................... 12 GeForce Experience Software Installation ...................................................................................... 12 05 CONFIGURING NVLink ......................................................................... 14 06 CONFIGURING HDMI ...........................................................................
    [Show full text]