Semiconductor the Dawn of Memory -Centric Computing

Total Page:16

File Type:pdf, Size:1020Kb

Semiconductor the Dawn of Memory -Centric Computing Semiconductor The dawn of memory -centric computing Demand: Memory-centric computing Neutral (Maintain) We recently attended Flash Memory Summit (FMS) 2017, held from August 7 th to 10 th in Santa Clara, California. Launched in 2006, FMS is the world’s largest Industry Report conference on NAND flash technology, and a number of milestone technologies, August 23, 2017 such as eMMC and 3D NAND, have previously been unveiled at the event. We noticed several important the mes at this year’s event. One key area of focus Mirae Asset Daewoo Co., Ltd. was memory-centric computing, a concept that places memory at the center of server architecture. Ever since the invention of computers, CPUs have been the [Semiconductor ] focal point, with memory playing a supporting role. H owever, the recent rise of new Hyunwoo Doh sources of demand that require high-performance memory (like machine learning) +822 -3774 -3803 has led to the notion that computer architecture should revolve around memory. [email protected] At this year’s FMS, there was much discussion of specific technol ogies for implementing memory-centric computing, with many presentations focusing on storage fabric technology. Currently, data centers operate at the individual server level, with data processed via each device’s CPUs, DRAM, and NAND. (The servers are linked through a network.) In contrast, memory-centric computing manages all data center server elements (CPU, DRAM, NAND, and HDD) in resource pools. CPUs are aggregated into one pool, and memory into another. This gives multiple CPUs parallel access to dat a stored in one massive memory pool, providing far better performance compared to data processing on CPUs and memory at the server level. Improved data center performance would fuel demand for machine learning, which, in turn, would boost overall semiconductor demand. Supply: QLCs, double stacking, and SCM While memory-centric computing was the biggest theme regarding memory demand, the main supply-side topics were quadruple-level cells (QLCs), double stacking, and storage-class memory (SCM). There was also widespread anticipation that 3D NAND capacity ramp-ups would pick up pace in 2018. QLCs, which store four bits per NAND cell, enable a 30% bit increase compared to the current triple- level cell (TLC) technology (based on size). While many previously believed QLCs would be difficult to implement because of reliability issues, this year’s conference shed light on advances in technology and pointed to the growing prospects of QLC commercialization. Given new technologies and capex plans, we believe NAND supply growth will begin to accelerate in 2018. In particular, the recent successful development of QLCs by Toshiba and Samsung Electronics (SEC; 005930 KS/Buy/TP: W2,950,000/CP: W2,350,000) has encouraged many engineers. QLC volume production is largely a nticipated to begin full swing in 2018. At this year’s event, we also heard commentary on how double stacking will facilitate the expansion of NAND supply. It was originally believed that scaling 3D NAND from the current 64 layers to 96 and beyond would gi ve rise to significant technological hurdles. However, there are increasing expectations that the development of double stacking technology, which involves stacking two separate devices, will make the expansion of layer count to 96 or above somewhat easier. Another widely discussed topic was SCM, a next-generation memory that sits between DRAM and NAND. The recent rapid advance of SCM has been driven by new sources of demand, such as machine learning. Because machine learning calls for huge amounts of data to be accessed quickly, DRAM is considered too expensive, while NAND is considered too slow. This is where SCM comes in, as it is more affordable than DRAM and faster than NAND. Many conference presenters highlighted the importance of SCM to the implementation of memory-centric computing. August 23, 2017 Semiconductor I. Flash Memory Summit 2017 We recently attended FMS 2017, held from August 7 th to 10 th in Santa Clara, California. Launched in 2006, FMS is the world’s largest conference on NAND flash technology, and a number of milestone technologies, such as eMMC, TLC, universal flash storage (UFS), 3D NAND, and 3D XPoint have previously been unveiled at the event. This year, engineers from leading NAND producers (SEC, SK Hynix [000660 KS/Trading Buy/TP: W75,000/CP: W68,200], Micron, Toshiba, etc.), controller producers (Marvell Technology Group, Silicon Motion, etc.), storage players (NetApp, Pure Storage, etc.), and internet giants (Google, Facebook, etc.) carried out seminars on new technologies. We noticed several important themes at this year’s event. One key area of focus was memory-centric computing, a concept that places memory at the center of server architecture. Ever since the invention of computers, CPUs have been the focal point, with memory playing a supporting role. However, the recent rise of new sources of demand that require high-performance memory (like machine learning) has led to the notion that computer architecture should revolve around memory. At this year’s FMS, there was much discussion of specific technologies for implementing memory-centric computing, with many presentations focusing on storage fabric technology. The subject matter of the FMS 2017 presentations suggests that the key driver of NAND demand has largely shifted from mobile to servers. Another widely discussed topic was SCM, including 3D XPoint and ReRAM, reflecting the rise of technologies requiring high- performance storage solutions like cloud computing and machine learning. In addition, there were a significant number of presentations covering non-volatile memory express (NVMe) and QLC technologies, as well as the outlook for the Chinese memory market. Flash memory for autonomous cars also attracted attention. We believe that the key takeaways from FMS 2017 were as follows: 1) NAND demand is anticipated to remain on the rise on the back of growth in the server market. Notably, memory-centric computing and storage fabric technologies are expected to drive the high-performance storage market. 2) Demand for SCM, including 3D XPoint and ReRAM, is projected to pick up in the near future. 3) NAND supply is expected to remain stable this year but should grow at an accelerated pace starting next year on the back of increasing 3D NAND capacity, Toshiba’s normalization, and the increasing adoption of new process technologies, including double stacking and QLC. 4) Chinese NAND producers will likely commence mass-production in 2H18 or 2019. While their technological competitiveness is believed to lag behind that of leading players like SEC, it should be on par with or surpass that of niche players. Mirae Asset Daewoo Research 2 August 23, 2017 Semiconductor Figure 1. Engineers from Facebook and Seagate present NVMe SSD controller technologies Source: FMS 2017, Mirae Asset Daewoo Research Figure 2. An eBay engineer explains the company’s data center structure Source: FMS 2017, Mirae Asset Daewoo Research Mirae Asset Daewoo Research 3 August 23, 2017 Semiconductor II. Demand: Memory-centric computing Data explosion was a common theme running through the presentations made at FMS 2017, suggesting that the profile of memory, particularly flash memory, will sharply rise amid the emergence of new sources of demand, such as machine learning, autonomous driving, and cloud computing. Significantly, these new demand drivers require high- performance memory as well as changes in server architecture/interfaces. Ever since the invention of computers, CPUs have been the focal point, with memory playing a supporting role. However, we are starting to see a shift toward the notion that computer architecture should revolve around memory. At this year’s FMS, there was much discussion of memory-centric computing technologies, with many presentations focusing on fabric computing technology. Currently, data centers operate at the individual server level, with data processed via each device’s CPUs, graphics processing units (GPUs), DRAM, HDDs and NAND. In contrast, fabric computing manages all data center server elements (CPUs, DRAM, NAND, and HDDs) in resource pools. CPUs are aggregated into one pool, and memory into another. This gives multiple CPUs parallel access to data stored in one massive memory pool, providing far better performance (particularly for machine learning, which requires parallel data processing), preventing computing resources from being wasted, and enabling flexible expansion/reduction in the number of severs per data center. However, in order to implement fabric computing, there are still many technological hurdles that need to be cleared. Notably, fabric computing can be divided into two types: static and dynamic. While the former can be implemented with existing technologies (as it is characterized by a hardware structure similar to that of conventional computing and the distribution of resources via software), the latter (which is fabric computing in its truest sense) is currently difficult to implement, as it requires real-time hardware division/integration. To implement dynamic fabric computing, it will be necessary to develop fresh networking technologies, as even the newest technologies, including HyperTransport and InfiniBand, have weaknesses. Dynamic fabric computing requires real-time integration of computing resources, which is impossible with existing technologies. Currently, linkages between CPUs and memory generally have response times of several nanoseconds
Recommended publications
  • Cisco Systems, Inc. 2014 Annual Report
    Cisco Systems, Inc. 2014 Annual Report Annual Report 2014 Letter to Shareholders To Our Shareholders, When we look back on fiscal 2014, we could describe to capture this opportunity. With this robust foundation, it as one of the most innovative years in Cisco’s history. we are able to provide our customers with an integrated, Through our innovation strategy of build, buy, partner, and network-centric architectural solution that not only solves integrate, we brought new architectures to market for the their IT challenges but also helps them achieve their desired next generation of networking, security, data center, and business outcomes. collaboration products and solutions, and we made bold Our vision is clear, and our strategy is working and has moves to help our customers capitalize on the changes in largely played out as we expected. In 2011, I said that the market. customers will view the network as the most strategic We managed our business well amid a great deal of change asset, not just in communications but also in IT and that this and a tough market. Our fiscal year began with several would enable us to become the number-one IT company. external headwinds, including the U.S. federal government I am more confident than ever that we can make that shutdown and the possibility of a U.S. default, combined aspiration a reality. with a significant slowdown in emerging countries. Even Innovation, Disruption, and Transformation with this backdrop, we ended fiscal 2014 with revenue of $47.1 billion, making it the second strongest year in our The pace of change being experienced in almost every history.
    [Show full text]
  • Impact of the New Generation of X86 on the Server Market
    Impact of the New Generation of x86 on the Server Market Gartner RAS Core Research Note G002012320, George J. Weiss, Jeffrey Hewitt, 3 June 2010, R3598 07042011 Intel and OEMs have driven performance and reliability of the Intel Xeon 7500 processor series (code-named Nehalem-EX) to a level overlapping what Gartner judges to be about 80% of the RISC/Itanium function/feature set, but at a much lower price point. The result will cause increased server technology consolidation toward a bipolar distribution: x86-based platforms and one primary market-share-leading RISC technology (IBM Power), complementing the prevailing legacy of mainframes. Key Findings • A bipolar technology distribution of one strong RISC (Power) and x86 will drive at least 70% to 80% of the market’s total revenue, while other technologies subsist with minor single-digit revenue share; beyond this decade, a unipolar market will exist in which x86 technology becomes the standard, predominant server technology. • Unix has receded by $4 billion in the past three years. Migrations from Unix to Linux or Windows continue, according to Gartner client discussions. • Users continue to optimize costs (such as through virtualization, lower hardware costs, open-source software and cloud computing) primarily on x86. • Of the three non-x86 technologies, IBM’s has had a recent track record of sustained market share growth. Recommendations • Calculate an approximate x86 total cost of ownership (TCO) equivalence or line item cost analysis for comparison as if these applications were to be hosted on x86 for similar performance/availability, when considering Unix upgrades. • Make the comparison to an approximately equivalent x86 configuration when evaluating price quotes.
    [Show full text]
  • 1 Cloud Computing the Present of Cloud Computing
    Outlines Cloud Computing: Past, Present, Introduction and Future The Past of Cloud Computing The Present of Cloud Computing The Future of Cloud Computing Ruay-Shiung Chang (張瑞雄) Department of Computer Science and Information Engineering Conclusions National Dong Hwa University (國立東華大學) November 4, 2010 “The distinction between past, present, and future is only a stubbornly persistent illusion” Albert Einstein 1 2 Introduction 3 4 1 Introduction Introduction Cloud computing in everything! Cloud computing=cloud+computing? There are many types of clouds! 5 6 Introduction Introduction Different angles have different opinions! The computer industry is the only industry that is more fashion-driven than women's fashion. ---Larry Ellision (Oracle CEO) Grid computing Cloud computing Utility computing Elastic comppguting Fabric computing Reconfigurable computing Social computing Volunteer computing SaaS (Software as a Service), PaaS, IaaS, … ASP (Application Service Provision) Data Center Co-location … 7 8 2 Introduction Introduction Oracle CEO talks about cloud computing Computer manufacturer Dell even tried to trademark the term "cloud computing", although its application was refused. But Dell got the domain name cloudcomputing.com, cloud-computing.com cloudcomputing.org (.net) is registered by r4l.com (待價而沽) cloud-computing.org (.net) (not used) thecloudcomputing.org (IEEE CC conference) 9 10 Introduction Introduction Who coined the phrase Cloud Computing? Who coined the phrase Cloud Computing? 11 12 3 Introduction Introduction 台灣翻譯成「雲端計算」 中華文化「雲端」的起源 大陸翻譯成「雲計算」 李白的詩「長相思二首之一」 那個高明? 長相思,在長安。絡緯秋啼金井闌, 有雲(servers)有端(access devices) 微霜淒淒簟色寒。孤燈不明思欲絕, 卷帷望月空長歎。美人如花隔雲端, 上有青冥之長天,下有淥水之波瀾。 天長地遠魂飛苦,夢魂不到關山難。 長相思,摧心肝。 We can’t see the cloud computing clearly yet. We can’t reach the cloud computing yet.
    [Show full text]
  • Mellanox Technologies, Arista Networks, Lawrence Livermore National Laboratory, Sun Microsystems and System Fabric Works Demonst
    Mellanox Technologies, Arista Networks, Lawrence Livermore National Laboratory, Sun Microsystems and System Fabric Works Demonstrate Lustre File System using Low Latency 10 Gigabit Ethernet at SC08 11/18/2008 Austin, Texas, Nov 18, 2008 - Mellanox® Technologies, Ltd. (NASDAQ: MLNX; TASE: MLNX), a leading supplier of semiconductor-based, server and storage interconnect products, today announced a multi-vendor 10 Gigabit Ethernet file system and storage infrastructure demonstration using a 10Gb/s Low Latency Ethernet (LLE) cluster network using Lustre™ technology for computational checkpointing and persistent visualization data storage. The demonstration includes the following vendor components: Mellanox - ConnectX EN 10 Gigabit Ethernet Adapters, Arista Networks - 7124 10 Gigabit Switches, Sun Microsystems - Lustre file system and Sun Fire™ X4540 storage server, System Fabric Works – software integration and demonstration deployment. “To maintain the cutting-edge R&D computing capabilities our scientists need to fulfill their national security missions, we look for the cost-effective technology solutions that best suit our unique computing environment,” said Mark Seager, Assistant Department Head for Advanced Technology of LLNL’s Computation Directorate. “Employing this kind of technology allows us to explore and develop the petascale computing capabilities vital to our future research endeavors and advances high performance computing.” “End-users are looking for improved application latency and performance for their compute intensive applications,
    [Show full text]
  • Lorem Ipsum Dolor Sit Amet Sit Consectetur Adipisicing Doloret HP
    1 Issue 1 3 From the Gartner Files: Magic Quadrant for Blade Servers Lorem ipsum dolor sit amet sit 17 About HP consecteturHP Blades Leadership adipisicing and doloret Innovation for the Data Center Featuring the Magic Quadrant for Blades Utilizing the best technology to solve your business needs is paramount to getting an edge on the competition, improving efficiency, and maximizing benefit from your IT budget. The Gartner Magic Quadrant is a trusted source of objective, independent intelligence on information technology. In Gartner’s recent blade server Magic “Due to their modular Quadrant, HP is positioned as a leader. nature, blades offer compelling facilities- Gartner Magic Quadrants are a culmination of oriented benefits, research in a specific market, giving you a wide- such as improved angle view of the relative positions of the market’s cabling, rapid hardware competitors. By applying a graphical treatment provisioning (including and a uniform set of evaluation criteria, a Gartner HP BladeSystem maximizes every hour, watt, and the ability to replace Magic Quadrant quickly helps you digest how well dollar in your systems—provisioning in minutes, failed components), technology providers are executing against their reducing equipment up to 95%, and slashing high computing stated vision. TCO up to 56%1. And our worldwide leadership density, energy- and consistent blade server innovation drives the efficient design and HP is positioned as a market leader for its “Ability change needed for IT to optimize business results. increasing management to Execute” and “Completeness of Vision”. We automation.” believe this solidifies HP’s leadership in blade HP is continually investing in breakthrough server technology, partnerships and solutions.
    [Show full text]
  • Magic Quadrant for Blade Servers
    Magic Quadrant for Blade Servers Gartner RAS Core Research Note G00207244, Andrew Butler, George J. Weiss, 17 January 2011, V3RA4 01202012 This Magic Quadrant for Blade Servers focuses on a market that is growing in complexity and sophistication, due to the emergence of related form factors, such as skinless servers, and the fast-growing interest in cloud computing, virtual input/output and fabrics. WHAT YOU NEED TO KNOW Blades represent an important stage in the evolution of servers as separate, discrete platforms give way to modular designs, and the boundaries between servers, storage and networking become increasingly blurred. This creates an increasing overlap of functionality between product categories that were previously more clearly delineated. These boundaries will be further tested as fabric-based computing becomes more prevalent, and all blade servers are examples of a fabric-enabled architecture, due to the switch-based backplane or midplane that they exploit. Blades represent a more proprietary investment, due to the lack of hardware form factor interoperability standards and the growing dependence on proprietary management tools and virtual input/output (I/O) solutions. Due to their modular nature, blades offer compelling operational benefits, such as improved cabling, rapid hardware provisioning, high compute density, energy-efficient design and increasing management automation. However, blades deliver few, if any, incremental application benefits, compared with their rack- and tower-based peers. Blades are not the only choice for modular deployment; rack-optimized servers deliver some modularity benefits, and skinless servers now represent an even more extreme class of high-density deployment. Because many vendors position their blade and skinless servers as part of a standard “modular server” portfolio, the nascent market status of skinless servers is reflected in this year’s blade server Magic Quadrant.
    [Show full text]
  • National Science Foundation Cri Workshop
    NATIONAL SCIENCE FOUNDATION Division of Computer and Network Systems CISE Computing Research Infrastructure CRI WORKSHOP SNOWBIRD UTAH JUNE 23-25, 2006 NATIONAL SCIENCE FOUNDATION Division of Computer and Network Systems CISE Computing Research Infrastructure CRI-WORKSHOP PROGRAM ORGANIZATION LOGISTICS Rita Rodriguez, CISE, NSF Dana Neill Steve Mahaney, CISE, NSF Computing Research Association Malek Adjouadi, FIU INVITED SPEAKERS Peter Freeman, Assistant Director CISE, NSF Deborah Crawford, Deputy Assistant Director, CISE, NSF PANELS CISE and the Global Implications: Larry Peterson, Princeton; David Novick, UTEP; Patrick Crowley, Washington U. St. Louis; Ram Dantu, U. of North Texas; Debby Cheng, Michigan State U.; and Bryant York, Portland State U. The GENI Program Guru Parulkar, GENI Program Director, CISE, NSF The Industry/University Cooperative Research Centers (I/UCRCs) Alex Schwarzkopf, Program Director, NSF, Division of Eng Education & Centers Sayfe Kiaei, Arizona State University, Director of Connection One and WINTech. ORGANIZATION OF PROCEEDINGS PHOTOS ON THE COVER PAGE Malek Adjouadi with thanks to the FIU-CATE Courtesy of the Cliff Lodge students Magno Guillen and Javier Delgado. Snowbird Ski & Summer Resort WITH THANKS To all of the investigators, presenters and participants of this NSF Workshop. SNOWBIRD UTAH JUNE 23-25, 2006 . NATIONAL SCIENCE FOUNDATION Division of Computer and Network Systems CISE Computing Research Infrastructure WORKSHOP - AGENDA Friday: June 23 Registration: 4:00 PM – 6:00PM Reception: 6:00 PM Dinner: 7:00 PM Saturday: June 24 7:30 - 8:00 Breakfast 8:00 - 8:20 Welcome Steve Mahaney, CISE/CNS Program Director, NSF Rita Rodriguez, CISE/CNS Program Director, NSF Malek Adjouadi, Florida International University 8:20 – 10:20 Session I: First-Year Awards Sabharwal, Ashutosh Rice University Shieber, Stuart Harvard University Callan, Jamie Carnegie Mellon University Manikopoulos, Constantine Foundation @ NJIT Chandra, Namas Florida A&M University Raju, G.V.S.
    [Show full text]
  • System on Fabrics Utilising Distributed Computing
    SYSTEM ON FABRICS UTILISING DISTRIBUTED COMPUTING by Partheepan Kandaswamy A Doctoral Thesis submitted in partial fulfilment of the requirements for the award of Doctor of Philosophy of Loughborough University June 2018 © by Partheepan Kandaswamy To my parents Shanthi and Kandaswamy ii ABSTRACT The main vision of wearable computing is to make electronic systems an important part of everyday clothing in the future which will serve as intelligent personal assistants. Wearable devices have the potential to be wearable computers and not mere input/output devices for the human body. The present thesis focuses on introducing a new wearable computing paradigm, where the processing elements are closely coupled with the sensors that are distributed using Instruction Systolic Array (ISA) architecture. The thesis describes a novel, multiple sensor, multiple processor system architecture prototype based on the Instruction Systolic Array paradigm for distributed computing on fabrics. The thesis introduces new programming model to implement the distributed computer on fabrics. The implementation of the concept has been validated using parallel algorithms. A real-time shape sensing and reconstruction application has been implemented on this architecture and has demonstrated a physical design for a wearable system based on the ISA concept constructed from off-the-shelf microcontrollers and sensors. Results demonstrate that the real time application executes on the prototype ISA implementation thus confirming the viability of the proposed architecture for fabric- resident computing devices. iii ACKNOWLEDGEMENTS I would like to thank my supervisors Dr. James Flint and Dr. Vassilios Chouliaras for their continuous support, guidance, exceptional advice and shared knowledge throughout my PhD. I would also thank Dr.
    [Show full text]
  • Ethernet Fabrics for the Cloud
    White Paper The Federal Perspective Ethernet Fabrics for the Cloud: “...use cloud- Setting the Stage for the Next-Generation based solutions Data Center whenever a An IDC White Paper by Lucinda Borovick, March 2011 In late 2010, the U.S. federal government announced its “Cloud First” policy. It requires secure, reliable, federal agencies to use cloud-based solutions “whenever a secure, reliable, cost- cost-effective effective cloud option exists.” Cloud computing provides highly reliable, scalable resources quickly, addressing a cloud option number of issues impacting the current government IT environment. However, today’s traditional Ethernet networks dominate existing data centers, and employ inefficient exists.” architectures that can be both complex to manage and costly to operate. Why Ethernet Fabrics? This white paper contends that next-generation data centers should employ an Ethernet fabric architecture. Specifically, it defines an Ethernet fabric as “a bigger, flatter network that preserves the low cost and simplicity of Ethernet, requires no reconfiguration or change in port configurations when resources move, and uses all paths available through the data center(s) for the most efficient transport of virtual machines (VMs) and their data.” The white paper explains in detail how Ethernet fabrics can better support the demands of the cloud and the IT resource demands of enterprises, which are remarkably similar in many ways to the needs of government agencies. Ethernet fabrics are: • Simpler to build and manage • Far more flexible
    [Show full text]
  • Computing Technologies and Practical Utility
    ISSN 2319-8885 Vol.04,Issue.37, September-2015, Pages:8054-8066 www.ijsetr.com Computing Technologies and Practical Utility K. J. SARMA SMIEEE, Professor, Dept of Humanities and Sciences and Academic Cell, Malla Reddy Engineering College (Autonomous), Secunderabad, TS, India, E-mail:[email protected]. Abstract: Several computing technologies have emerged by the end of 20 th century due to the high speed Broadband internet with each having a number of applications in science, engineering, business, social systems, and governments. These technologies brought revolutionary changes in communications and computational processes. In- fact the advances in hardware, investigation of sophisticated rigorous mathematical algorithms are also responsible for greater shift in these developments. Some of these computing technologies are cluster, grid, and cloud, cloud-mobile. Multi-cloud, parallel, concurrent, distributed, DNA, mobile, high performance, utility, global cooperative, functional, ubiquitous, pervasive, exploratory, quantum, scalable computing, ubiquitous secure, computational grid, etc. In this paper we review uses and practical applications. Research into enhancing the technology, utility in various application areas have been growing at fast phase. It may stimulate further study of each and the convergence of all these technologies in dealing with logistic based problems of large dimensionality. Keywords: Cluster, Grid, Cloud, Distributed, DNA, Mobile, High Performance, Ubiquitous Pervasive, Quantum. I. INTRODUCTION computing clusters. Desktop work-stations which can also Cluster computing is the topic of research among become part of a cluster when they’re not being used. academics, industry community, system designers, network Financial services firm, which probably has many high- developers, language designers, technical forums, algorithm powered workstations that sit idle overnight.
    [Show full text]
  • Emerging Fabric Technologies Technology Brief
    Emerging Fabric Technologies technology brief Abstract.............................................................................................................................................. 2 Introduction......................................................................................................................................... 2 Data center requirements for fabric technologies ..................................................................................... 2 Network interconnect challenges ........................................................................................................... 3 Key enabling technologies for fabrics..................................................................................................... 3 TCP/IP Offload Engines .................................................................................................................... 3 RDMA technology ............................................................................................................................ 4 HP advances RDMA and TOE technologies......................................................................................... 4 Emerging fabric technologies ................................................................................................................ 4 RDMA/TCP ..................................................................................................................................... 4 How RDMA/TCP works................................................................................................................
    [Show full text]
  • Windows Azure Platform
    Overview of Cloud Computing and the Windows Azure Platform Krishna Kumar [email protected] http://azurepilot.com Twitter: KrishnaOnAzure http://AzurePilot.com Evolution of computing paradigms + – 1960s Mainframes ruled the earth. sw dx + – 1970s The rise of the Minis sw dx + – 1980s Proliferation of the PCs sw dx + – 1990s Network the Servers sw dx + – 2000s The Internet comes of age sw dx Platform Evolution • Client HostedServer software platformMobile Cloud • Shared infrastructure • Virtualized and dynamic • Increasingly higher level services • Pay as you go pricing model + – 2010s Welcome to the Cloud sw dx Platform Continuum On-Premises Hosted Cloud Servers Servers Platform • Bring your own machines, • Renting machines, • Shared, multi-tenant connectivity, software, connectivity, software infrastructure etc. • Less control • Virtualized & dynamic • Complete control • Fewer responsibilities • Scalable & available • Complete responsibility • Lower capital costs • Abstracted from the • Static capabilities • More flexible infrastructure • Upfront capital costs for • Pay for fixed capacity, • Higher-level services the infrastructure even if idle • Pay as you go On-Premise % of Capital Equipment Budget spent on IT in 2000? 45% Commerce Department Statistics % of Utilized Server Capacity on Average? 6% Economist Survey on IT, 2008 Elasticity – Provisioning for Peak Real World Server Utilization Is 5% to 20% • Many Services Peak Exceeds Average by a Factor of 2 to 10 • Most Provision for Peak • Painful to Under-Provision (Lost Customers) Provisioning for Peak Without Elasticity, We Waste Resources (Shaded Areas) During Non-Peak Times Elasticity: Risks of Under-Provisioning Under-Provisioning #1 Potential Revenue (Shaded Area) Is Sacrificed Under-Provisioning #2 Some Users Respond to Under-Provisioning by Permanently Deserting the Site..
    [Show full text]