Performance of Various Computers Using Standard Linear Equations

Total Page:16

File Type:pdf, Size:1020Kb

Performance of Various Computers Using Standard Linear Equations CS Performance of Various Computers Using Standard Linear Equations Software Jack J Dongarra Computer Science Department UniversityofTennessee Knoxville TN and Mathematical Sciences Section Oak Ridge National Lab oratory Oak Ridge TN CS Novemb er Electronic mail address dongarracsutkedu An uptodate version of this rep ort can b e found at httpwwwnetliborgb enchmarkp erformanceps This work was supp orted in part by the Applied Mathematical Sciences subprogram of the Oce of Energy Research US Department of Energy under Contract DEACOR and in part by the Science Alliance a state supp orted program at the UniversityofTennessee Performance of Various Computers Using Standard Linear Equations Software Jack J Dongarra Computer Science Department UniversityofTennessee Knoxville TN and Mathematical Sciences Section Oak Ridge National Lab oratory Oak Ridge TN Novemb er Abstract This rep ort compares the p erformance of dierent computer systems in solving dense systems of linear equations The comparison involves approximately a hundred computers ranging from aCray YMP to scientic workstations such as the Ap ollo and Sun to IBM PCs Intro duction and Ob jectives The timing information presented here should in no way b e used to judge the overall p erformance of a computer system The results reect only one problem area solving dense systems of equations This rep ort provides p erformance information on a wide assortment of computers ranging from the homeused PC up to the most powerful sup ercomputers The information has been collected over a period of time and will undergo change as new machines are added and as hardware and software systems improve The programs used to generate this data can easily be obtained over the Internet While wemakeevery attempt to verify the results obtained from users and vendors errors are b ound to exist and should b e brought to our attention We encourage users to obtain the programs and run the routines on their machines rep orting any discrepancies with the numb ers listed here The rst table rep orts three numbers for each machine listed in some cases the numb ers are missing b ecause of lack of data All p erformance numbers reect arithmetic p erformed in full precision usually bit unless noted On some machines full precision may b e single precision such as the Cray or double precision suchasthe IBM The rst numb er is for the LINPACK b enchmark program for a matrix of order in a Fortran environment The second numb er is for solving a system of equations of order with no restriction on the metho d or its implementation The third numb er is the theoretical p eak p erformance of the machine LINPACK programs can be characterized as having a high p ercentage of oatingp oint arith metic op erations The routines involved in this timing study SGEFA and SGESL use column oriented algorithms That is the programs usually reference array elements sequentially down a This work was supp orted in part by the Applied Mathematical Sciences subprogram of the Oce of Energy Research US Department of Energy under Contract DEACOR and in part by the Science Alliance a state supp orted program at the UniversityofTennessee An uptodate version of this rep ort can b e found at httpwwwnetliborgb enchmarkp erformanceps November column not across a row Column orientation is imp ortant in increasing eciency b ecause of the wayFortran stores arrays Most oatingp oint op erations in LINPACK take place in a set of subpro grams the Basic Linear Algebra Subprograms BLAS which are called rep eatedly throughout the calculation These BLAS referred to nowasLevel BLAS reference onedimensional arrays rather than twodimensional arrays In the rst case the problem size is relatively small order and no changes were made to the LINPACK software Moreover no attempt was made to use sp ecial hardware features or to exploit vector capabilities or multiple pro cessors The compilers on some machines may of course generate optimized co de that itself accesses sp ecial features Thus many highp erformance machines maynothavereached their asymptotic execution rates In the second case the problem size is larger matrix of order and mo difying or replac ing the algorithm and software was p ermitted to achieve as high an execution rate as p ossible Thus the hardware had more opp ortunity for reaching nearasymptotic rates An imp ortantcon straint however was that all optimized programs maintain the same relative accuracy as standard techniques such as Gaussian elimination used in LINPACK Furthermore the driver program supplied with the LINPACK benchmark had to be run to ensure that the same problem is solved The driver program sets up the matrix calls the routines to solve the problem veries that the answers are correct and computes the total number of op erations to solve the problem indep endent of the metho d as n n wheren The last column is based not on an actual program run but on a pap er computation to determine the theoretical p eak Mops rate for the machine This is the number manufacturers often cite it represents an upp er b ound on p erformance That is the manufacturer guarantees that programs will not exceed this ratesort of a sp eed of light for a given computer The theoretical p eak p erformance is determined by counting the number of oatingp oint ad ditions and multiplications in full precision that can be completed during a p erio d of time usually the cycle time of the machine As an example the Cray YMP has a cycle time of ns During a cycle the results of both an addition and a multiplication can be completed op erations cycle Mops on a single pro cessor On the Cray YMP there are ns cycle pro cessors thus the p eak p erformance is Mops The information in this rep ort is presented to users to provide a range of p erformance for the various computers and to show the eects of typical Fortran programming and the results that can b e obtained through careful programming The maximum rate of execution is given for comparison The column lab eled Computer gives the name of the computer hardware on which the program was run In some cases we have indicated the number of pro cessors in the conguration and in some cases the cycle time of the pro cessor in nanoseconds The column lab eled LINPACK Benchmark gives the op erating system and compiler used The run was based on two routines from LINPACK SGEFA and SGESL were used for single precision and DGEFA and DGESL were used for double precision These routines p erform standard LU decomp osition with partial pivoting and backsubstitution The timing was done on a matrix of order where no changes are allowed to the Fortran programs The column lab eled TPP Toward Peak Performance gives the results of hand optimization the problem size was of order The nal column lab eled Theoretical Peak gives the maximum rate of execution based on the cycle time of the hardware The same matrix was used to solve the system of equations The results were checked for accuracy by calculating a residual for the problem jjAx bjjjjAjjjjxjj The term Mops used as a rate of execution stands for millions of oatingp oint op erations completed p er second For solving a system of n equations n n op erations are p erformed November wecount b oth additions and multiplications The information in the tables was compiled over a p erio d of time Subsequent systems software and hardware changes may alter the timings to some extent One further note The following tables should not b e taken to o seriously In multiprogramming environments it is often dicult to reliably measure the execution time of a single program We trust that anyone actually evaluating machines and op erating systems will gather more reliable and more representativedata A Lo ok at Parallel Pro cessing While collecting the data presented in Table wewere able to exp eriment with parallel pro cessing onanumb er of computer systems For these exp eriments we used either the standard LINPACK algorithm or an algorithm based on matrixmatrix techniques In the case of the LINPACK algorithm the lo op around the SAXPY can b e p erformed in parallel In the matrixmatrix imple mentation the matrix pro duct can be split into submatrices and p erformed in parallel In either case the parallelism follows a simple forkandjoin mo del where each pro cessor gets some number of op erations to p erform For a problem of size we exp ect a high degree of parallelism Thus it is not surprising that we get such high eciency see Table The actual p ercentage of parallelism of course dep ends on the algorithm and on the sp eed of the unipro cessor on the parallel part relative to the sp eed of the unipro cessor on the nonparallel part Highly Parallel Computing With the arrival of massively parallel computers there is a need to benchmark such machines on problems that make sense The problem size and rule for the runs reected in the Tables and do not p ermit massively parallel computers to demonstrate their p otential p erformance The basic aw is the problem size is to o small Toprovide a forum for comparing suchmachines the following b enchmark was run on a numb er of massively parallel machines The b enchmark involves solving a system of linear equations as was done in Tables and However in this case the problem size is allowed to increase and the p erformance numb ers reect the largest problem run on the machine The ground rules are as follows Solve systems of linear equations bysome metho d allow the size of the problem to vary and measure the execution
Recommended publications
  • Sun Blade 6000 Chassis, the Foundation Equivalent Rackmount Servers of the Revolutionary Sun Blade 6000 Modular System
    Sun Blade™ 6000 Chassis The industry’s most open and versatile enterprise blade platform Highlights • Up to double the memory and Your datacenter is pressured by recent trends and constraints — the move to I/O capacity of competing blades multicore processors is driving increasingly larger node sizes; the maxing out of and rackmounts > power and cooling resources; the growing complexity of management; the needs for • Shared power and cooling — virtualization and consolidation; and the ongoing needs for expansion, repairs, and save 15 percent more energy than upgrades. Sun has a new answer to your problems: the Sun Blade 6000 Chassis, the foundation equivalent rackmount servers of the revolutionary Sun Blade 6000 modular system. The chassis is designed to deliver radical • Save up to $700,000 per rack in efficiencies to your datacenter by providing maximum performance packed into optimal cool- acquisition costs ing, space, and power requirements. It also supports a wider variety of applications, offering • Maximum flexibility – SPARC, AMD blades based on UltraSPARC®, AMD Opteron, and Intel Xeon processors. Opteron, and Intel Xeon proces- sor-based server modules, as well as the Solaris, Linux, and Windows The Sun Blade 6000 Chassis reduces costs possible with traditional blade computing. operating systems and provides superior performance, maximiz- It keeps costs down if you are virtualizing or • Most open blade platform with in- ing return on investment (ROI). The chassis building out large-scale HPC clusters or dustry-standard PCIe ExpressModule fits in a compact form factor — only 10U — Web farms, and helps take new systems to • Streamlined, Sun Blade Transparent while supporting up to 10 full-featured, production faster and at reduced cost.
    [Show full text]
  • Sun Ultratm 25 Workstation & Sun Ultra 45 Workstation Just the Facts
    Sun UltraTM 25 Workstation & Sun Ultra 45 Workstation Just the Facts SunWIN Token# 473547 SunWIN Token# 460409 Copyrights © 2006 Sun Microsystems, Inc. All Rights Reserved. Sun, Sun Microsystems, the Sun logo, Ultra, Sun Blade, Java, Solaris, Java, NetBeans, Sun Fire, Sun StorEdge, SunLink, SunSpectrum, SunSpectrum Platinum, SunSpectrum Gold, SunSpectrum Silver, SunSpectrum Bronze, SunSolve, SunPCi, and SunVTS are trademarks or registered trademarks of Sun Microsystems, Inc. in the United States and other countries. All SPARC trademarks are used under license and are trademarks or registered trademarks of SPARC International, Inc. in the United States and other countries. Products bearing SPARC trademarks are based upon an architecture developed by Sun Microsystems, Inc. UNIX is a registered trademark in the United States and other countries, exclusively licensed through X/Open Company, Ltd. Ultra 25/45 JTF - 12/10/07 Sun Confidential – Internal Use Only 2 Table of Contents Positioning.....................................................................................................................................................................4 Introduction...............................................................................................................................................................4 Product Family Placement .......................................................................................................................................5 Sun Ultra 45 vs Sun Ultra 25 Workstation...............................................................................................................5
    [Show full text]
  • 3Dfx Oral History Panel Gordon Campbell, Scott Sellers, Ross Q. Smith, and Gary M. Tarolli
    3dfx Oral History Panel Gordon Campbell, Scott Sellers, Ross Q. Smith, and Gary M. Tarolli Interviewed by: Shayne Hodge Recorded: July 29, 2013 Mountain View, California CHM Reference number: X6887.2013 © 2013 Computer History Museum 3dfx Oral History Panel Shayne Hodge: OK. My name is Shayne Hodge. This is July 29, 2013 at the afternoon in the Computer History Museum. We have with us today the founders of 3dfx, a graphics company from the 1990s of considerable influence. From left to right on the camera-- I'll let you guys introduce yourselves. Gary Tarolli: I'm Gary Tarolli. Scott Sellers: I'm Scott Sellers. Ross Smith: Ross Smith. Gordon Campbell: And Gordon Campbell. Hodge: And so why don't each of you take about a minute or two and describe your lives roughly up to the point where you need to say 3dfx to continue describing them. Tarolli: All right. Where do you want us to start? Hodge: Birth. Tarolli: Birth. Oh, born in New York, grew up in rural New York. Had a pretty uneventful childhood, but excelled at math and science. So I went to school for math at RPI [Rensselaer Polytechnic Institute] in Troy, New York. And there is where I met my first computer, a good old IBM mainframe that we were just talking about before [this taping], with punch cards. So I wrote my first computer program there and sort of fell in love with computer. So I became a computer scientist really. So I took all their computer science courses, went on to Caltech for VLSI engineering, which is where I met some people that influenced my career life afterwards.
    [Show full text]
  • PACKET 7 BOOKSTORE 433 Lecture 5 Dr W IBM OVERVIEW
    “PROCESSORS” and multi-processors Excerpt from Hennessey Computer Architecture book; edits by JT Wunderlich PhD Plus Dr W’s IBM Research & Development: JT Wunderlich PhD “PROCESSORS” Excerpt from Hennessey Computer Architecture book; edits by JT Wunderlich PhD Historical Perspective and Further 7.14 Reading There is a tremendous amount of history in multiprocessors; in this section we divide our discussion by both time period and architecture. We start with the SIMD SIMD=SinGle approach and the Illiac IV. We then turn to a short discussion of some other early experimental multiprocessors and progress to a discussion of some of the great Instruction, debates in parallel processing. Next we discuss the historical roots of the present multiprocessors and conclude by discussing recent advances. Multiple Data SIMD Computers: Attractive Idea, Many Attempts, No Lasting Successes The cost of a general multiprocessor is, however, very high and further design options were considered which would decrease the cost without seriously degrading the power or efficiency of the system. The options consist of recentralizing one of the three major components. Centralizing the [control unit] gives rise to the basic organization of [an] . array processor such as the Illiac IV. Bouknight, et al.[1972] The SIMD model was one of the earliest models of parallel computing, dating back to the first large-scale multiprocessor, the Illiac IV. The key idea in that multiprocessor, as in more recent SIMD multiprocessors, is to have a single instruc- tion that operates on many data items at once, using many functional units (see Figure 7.14.1). Although successful in pushing several technologies that proved useful in later projects, it failed as a computer.
    [Show full text]
  • Sun Blade 1000 and 2000 Workstations
    Sun BladeTM 1000 and 2000 Workstations Just the Facts Copyrights 2002 Sun Microsystems, Inc. All Rights Reserved. Sun, Sun Microsystems, the Sun logo, Sun Blade, PGX, Solaris, Ultra, Sun Enterprise, Starfire, SunPCi, Forte, VIS, XGL, XIL, Java, Java 3D, SunVideo, SunVideo Plus, Sun StorEdge, SunMicrophone, SunVTS, Solstice, Solstice AdminTools, Solstice Enterprise Agents, ShowMe, ShowMe How, ShowMe TV, Sun Workstation, StarOffice, iPlanet, Solaris Resource Manager, Java 2D, OpenWindows, SunCD, Sun Quad FastEthernet, SunFDDI, SunATM, SunCamera, SunForum, PGX32, SunSpectrum, SunSpectrum Platinum, SunSpectrum Gold, SunSpectrum Silver, SunSpectrum Bronze, SunSolve, SunSolve EarlyNotifier, and SunClient are trademarks, registered trademarks, or service marks of Sun Microsystems, Inc. in the United States and other countries. All SPARC trademarks are used under license and are trademarks or registered trademarks of SPARC International, Inc. in the United States and other countries. Products bearing SPARC trademarks are based upon an architecture developed by Sun Microsystems, Inc. UNIX is a registered trademark in the United States and in other countries, exclusively licensed through X/Open Company, Ltd. FireWire is a registered trademark of Apple Computer, Inc., used under license. OpenGL is a trademark of Silicon Graphics, Inc., which may be registered in certain jurisdictions. Netscape is a trademark of Netscape Communications Corporation. PostScript and Display PostScript are trademarks of Adobe Systems, Inc., which may be registered in
    [Show full text]
  • ADAC® a Philips Medical Systems Company Copyright Statement
    DICOM 3.0 CONFORMANCE STATEMENT Pegasys 9605-0132 Rev. C 11-2002 ADAC® A Philips Medical Systems Company Copyright Statement Philips Medical Systems has taken care to ensure the accuracy of this document. However, Philips Medical Systems assumes no liability for errors or omissions and reserves the right to make changes without further notice to any products herein to improve reliability, function, or design. Philips Medical Systems provides this guide without warranty of any kind, either implied or expressed, including, but not limited to, the implied warranties of merchantability and fitness for a particular purpose. Philips Medical Systems may make improvements or changes in the product(s) and/or program(s) described in this manual at any time. This document contains proprietary information which is protected by copyright. All rights are reserved. No part of this manual may be photocopied, reproduced, or translated to another language without written permission from Philips Medical Systems. Philips Medical Systems reserves the right to revise this publication and to make changes in content from time to time without obligation on the part of Philips Medical Systems to provide notification of such revision or change. Disclaimer Philips Medical Systems DICOM software is in compliance with the ACR-NEMA DICOM 3.0 standard; however, due to the inherent nature of DICOM, the user must perform acceptance testing to verify that the Philips Medical Systems DICOM software meets the requirements for your configuration. The acceptance testing must include all representative datasets (images) that you intend to transfer, all types of transfers desired for a type of dataset, and clinical evaluation of each representative dataset on the receiving end after each desired type of transfer.
    [Show full text]
  • Online Sec 6.15.Indd
    6.155.9 Historical Perspective and Further Reading Th ere is a tremendous amount of history in multiprocessors; in this section we divide our discussion by both time period and architecture. We start with the SIMD approach and the Illiac IV. We then turn to a short discussion of some other early experimental multiprocessors and progress to a discussion of some of the great debates in parallel processing. Next we discuss the historical roots of the present multiprocessors and conclude by discussing recent advances. SIMD Computers: Attractive Idea, Many Attempts, No Lasting Successes Th e cost of a general multiprocessor is, however, very high and further design options were considered which would decrease the cost without seriously degrading the power or effi ciency of the system. Th e options consist of recentralizing one of the three major components. Centralizing the [control unit] gives rise to the basic organization of [an] . array processor such as the Illiac IV. Bouknight et al. [1972] Th e SIMD model was one of the earliest models of parallel computing, dating back to the fi rst large-scale multiprocessor, the Illiac IV. Th e key idea in that multiprocessor, as in more recent SIMD multiprocessors, is to have a single instruction that operates on many data items at once, using many functional units (see Figure 6.15.1). Although successful in pushing several technologies that proved useful in later projects, it failed as a computer. Costs escalated from the $8 million estimate in 1966 to $31 million by 1972, despite construction of only a quarter of the planned multiprocessor.
    [Show full text]
  • Services for Oracle Sun Blade Servers
    ORACLE DATA SHEET SERVICES FOR ORACLE SUN BLADE SERVERS Oracle's Sun Blade modular systems integrate x86 and SPARC-based servers, INSTALL, CONFIGURE, AND OPTIMIZE ORACLE SUN BLADE SERVERS storage, and advanced networking capabilities to support a complex and dynamic mix of IT workloads. And with Oracle Advanced Customer Support KEY FEATURES Services you get the guidance you need to plan, deploy and keep your Oracle • Technical knowledge transfer tailored to Sun Blade technology optimized for peak performance. your IT environment and business needs With intimate knowledge of Oracle tools and best practices, Oracle Advanced • Comprehensive, system hardware installation including site audit, installation Customer Support Services provides the right knowledge at the right time to help and configuration planning documentation, mitigate risk and to maximize the value of your Oracle technology investment. testing and handover • Trusted systems knowledge and proven IT delivery methodologies Pre-production Services for Oracle Sun Blade Servers • Reviews and recommendations for Oracle Advanced Customer Support Services delivers optimization services that help you optimized deployment planning, system install, configure, optimize and support your Oracle technology environment. Whether you are configuration and meeting availability requirements initiating a technology refresh project, optimizing an existing environment through the use of virtualization technologies, or looking to drive better datacenter standardization and operational best practices,
    [Show full text]
  • Sun Blade™ 150 Workstation
    Datasheet Sun Blade 150 Workstation On the Web http://www.sun.com/sunblade150 Sun Blade™ 150 Workstation The price leader in 64-bit UNIX® workstations. Key feature highlights Entry-level workstation designed with full features to deliver exceptional performance. • 550-MHz or 650-MHz UltraSPARC® IIi For technical and business computing environments looking to balance advanced 64-bit capabili- processor with 512 KB L2 on-die cache ties, workstation performance, support for the Solaris Operating System, and cost considerations, • Up to 2 GB of SDRAM memory with Error Sun has the ideal solution: The Sun Blade™ 150 workstation. This powerful, low-cost system boasts Correction a full 64-bit architecture with the latest high-performance UltraSPARC IIi CPU, supporting up to 2 • Up to two ATA-66 80-GB, 7200-rpm hard GB of memory and up to 160 GB of internal storage. Features such as a fast PCI bus and the latest disk drives I/O technologies including USB and IEEE-1394 provide connectivity to a wide variety of latest- • 1394a (FireWire) generation peripherals. And with graphics options that include the high-resolution, on-board Sun™ • Three PCI slots PGX64 2-D, Sun™ XVR-100 2-D, or high-end Sun™ XVR-500 3-D graphics, the Sun Blade 150 worksta- • 10/100 BaseT Ethernet tion delivers advanced workstation capabilities with exceptional performance at an affordable • Solaris™ 8 HW 5/03 and Solaris 9 HW 4/03 price. With only a 20 percent increase in megahertz, the doubling of the cache to 512 KB, plus Sun Operating System pre-installed XVR-500 graphics, the Sun Blade 150 workstation cranks up the performance.
    [Show full text]
  • Nineteenth Status Report on Implementation
    WORLD METEOROLOGICAL ORGANIZATION WORLD WEATHER WATCH NINETEENTH STATUS REPORT ON IMPLEMENTATION 1999 WMO-No. 894 Secretariat of the World Meteorological Organization - Geneva - Switzerland WORLD METEOROLOGICAL ORGANIZATION WORLD WEATHER WATCH NINETEENTH STATUS REPORT ON IMPLEMENTATION 1999 WMO-No.894 Secretariat of the World Meteorological Organization- Geneva- Switzerland © 1999, World Meteorological Organization ISBN 92- 63- 10894- 3 NOTE The designations employed and the presentation of material in this publication do not imply the expression of any opinion whatsoever on the part of the Secretariat of the World Meteorological Organizati on concerning the legal status of any country, terri tory, city or area, of its authorities, or concern ing the delimitation of its frontiers or boundaries. FOREWORD The World Weather Watch (WWW) Programme of the World Meteorological Organization (WMO) was adopted in 1963 by the Fourth World Meteorological Congress. Since then it has grown to be fundamental for all WMO and other related programmes as regards the provision of basic meteorological data and products, telecommunication services and the management thereof. In continuing to attribute the highest priority to the furtherance and implementation of WWW, Twelfth Congress in Resolution 2 urged all Members of the Organization to cooperate actively and enthusiastically in the implementation of the WWW. This publication is the 19th in a series of biannual reports on the status of implementation of WWW. It was mainly designed to inform the senior management of the national Meteorological and Hydrological Services (NMHSs), but also interested academia and the private sector of the operational status of WWW. It provides information concerning the structure, status and trends of the implementation, as well as performance of the core components of WWW, notably the Global Observing System (GOS), the Global Telecommunication System (GTS) and the Global Data-processing System (GDPS).
    [Show full text]
  • AVS on UNIX WORKSTATIONS INSTALLATION/ RELEASE NOTES
    _________ ____ AVS on UNIX WORKSTATIONS INSTALLATION/ RELEASE NOTES ____________ Release 5.5 Final (50.86 / 50.88) November, 1999 Advanced Visual Systems Inc.________ Part Number: 330-0120-02 Rev L NOTICE This document, and the software and other products described or referenced in it, are con®dential and proprietary products of Advanced Visual Systems Inc. or its licensors. They are provided under, and are subject to, the terms and conditions of a written license agreement between Advanced Visual Systems and its customer, and may not be transferred, disclosed or otherwise provided to third parties, unless oth- erwise permitted by that agreement. NO REPRESENTATION OR OTHER AFFIRMATION OF FACT CONTAINED IN THIS DOCUMENT, INCLUDING WITHOUT LIMITATION STATEMENTS REGARDING CAPACITY, PERFORMANCE, OR SUI- TABILITY FOR USE OF SOFTWARE DESCRIBED HEREIN, SHALL BE DEEMED TO BE A WARRANTY BY ADVANCED VISUAL SYSTEMS FOR ANY PURPOSE OR GIVE RISE TO ANY LIABILITY OF ADVANCED VISUAL SYSTEMS WHATSOEVER. ADVANCED VISUAL SYSTEMS MAKES NO WAR- RANTY OF ANY KIND IN OR WITH REGARD TO THIS DOCUMENT, INCLUDING BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PUR- POSE. ADVANCED VISUAL SYSTEMS SHALL NOT BE RESPONSIBLE FOR ANY ERRORS THAT MAY APPEAR IN THIS DOCUMENT AND SHALL NOT BE LIABLE FOR ANY DAMAGES, INCLUDING WITHOUT LIMI- TATION INCIDENTAL, INDIRECT, SPECIAL OR CONSEQUENTIAL DAMAGES, ARISING OUT OF OR RELATED TO THIS DOCUMENT OR THE INFORMATION CONTAINED IN IT, EVEN IF ADVANCED VISUAL SYSTEMS HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH DAMAGES. The speci®cations and other information contained in this document for some purposes may not be com- plete, current or correct, and are subject to change without notice.
    [Show full text]
  • Hive: Fault Containment for Shared-Memory Multiprocessors
    Hive: Fault Containment for Shared-Memory Multiprocessors John Chapin, Mendel Rosenblum, Scott Devine, Tirthankar Lahiri, Dan Teodosiu, and Anoop Gupta Computer Systems Laboratory Stanford University, Stanford CA 94305 htkp: //www-flash. skanforcl. edu Abstract In this paper we describe Hive, an operahng system designed for large-scale shared-memory multiprocessors. Hive is Reliability and scalability are major concerns when designing fundamentally different from previous monolithic and microkernel operating systems for large-scale shared-memory multiprocessors. SMP OS implementations: it IS structured as an internal distributed In this paper we describe Hive, an operating system with a novel system of independent kernels called ce/ls. This multicellular kernel architecture that addresses these issues Hive is structured kernel architecture has two main advantages: as an internal distributed system of independent kernels called ● Re[mbiltry: In SMP OS implementations, any significant cells. This improves reliabihty because a hardwme or software hardware or software fault causes the entire system to crash. fault damages only one cell rather than the whole system, and For large-scale machines this can result in an unacceptably low improves scalability because few kernel resources are shared by mean time to failure. In Hive, only the cell where the fault processes running on different cells. The Hive prototype is a occurred crashes, so only the processes using the resources of complete implementation of UNIX SVR4 and is targeted to run on that cell are affected. This is especially beneficial for compute the Stanford FLASH multiprocessor. server workloads where there are multiple independent This paper focuses on Hive’s solutlon to the following key processes.
    [Show full text]