SGI® Fibre Channel PCI Option Board and XIO™ Option Board User's
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Paper in PDF Format
Managing System Resources on ‘Teras’ Experiences in Managing System Resources on a Large SGI Origin 3800 Cluster The paper to the presentation given at the CUG SUMMIT 2002 held at the University of Manchester. Manchester (UK), May 20-24 2002. http://www.sara.nl Mark van de Sanden [email protected] Huub Stoffers [email protected] Introduction Outline of the Paper This is a paper about our experiences with managing system resources on ‘Teras’. Since November 2000 Teras, a large SGI Origin 3800 cluster, is the national supercomputer for the Dutch academic community. Teras is located at and maintained by SARA High Performance Computing (HPC). The outline of this presentation is as follows: Very briefly, we will tell you what SARA is, and what we presently do, particularly in high performance computing, besides maintaining Teras. Subsequently, to give you an idea, on the one hand which resources are available on Teras, and on the other hand what resources are demanded, we will give overviews of three particular aspects of the Teras: 1) Some data on the hardware – number of CPUs, memory, etc. - and the cluster configuration, i.e. description of the hosts that constitute the cluster. 2) Identification of the key software components that are used on the system; both, system software components, as well as software toolkits available to users to create their (parallel) applications. 3) A characterization of the ‘job mix’ that runs on the Teras cluster. Thus equipped with a clearer idea of both ‘supply’ and ‘demand’ of resources on Teras, we then state the resource allocation policy goals that we pursue in this context, and we review what resource management possibilities we have at hand and what they can accomplish for us, in principle. -
Overview and History Nas Overview
zjjvojvovo OVERVIEWNAS OVERVIEW AND HISTORY a Look at NAS’ 25 Years of Innovation and a Glimpse at the Future In the mid-1970s, a group of Ames aerospace researchers began to study a highly innovative concept: NASA could transform U.S. aerospace R&D from the costly and cghghmgm time-consuming wind tunnel-based process to simulation-centric design and engineer- ing by executing emerging computational fluid dynamics (CFD) models on supercom- puters at least 1,000 times more powerful than those commercially available at the time. In 1976, Ames Center Director Dr. Hans Mark tasked a group led by Dr. F. Ronald Bailey to explore this concept, leading to formation of the Numerical Aerodynamic Simulator (NAS) Projects Office in 1979. At the same time, a user interface group was formed consisting of CFD leaders from industry, government, and academia to help guide requirements for the NAS concept gmfgfgmfand provide feedback on evolving computer feasibility studies. At the conclusion of these activities in 1982, NASA management changed the NAS approach from a focus on purchasing a specially developed supercomputer to an on-going Numerical Aerody- namic Simulation Program to provide leading-edge computational capabilities based on an innovative network-centric environment. The NAS Program plan for implementing this new approach was signed on February 8, 1983. Grand Opening As the NAS Program got underway, a projects office to a full-fledged division its first supercomputers were installed at Ames. In January 1987, NAS staff and in Ames’ Central Computing Facility, equipment were relocated to the new starting with a Cray X-MP-12 in 1984. -
SGI™ Origin™ 200 Scalable Multiprocessing Server Origin 200—In Partnership with You
Product Guide SGI™ Origin™ 200 Scalable Multiprocessing Server Origin 200—In Partnership with You Today’s business climate requires servers that manage, serve, and support an ever-increasing number of clients and applications in a rapidly changing environment. Whether you use your server to enhance your presence on the Web, support a local workgroup or department, complete dedicated computation or analy- sis, or act as a core piece of your information management infrastructure, the Origin 200 server from SGI was designed to meet your needs and exceed your expectations. With pricing that starts on par with PC servers and performance that outstrips its competition, Origin 200 makes perfect business sense. •The choice among several Origin 200 models allows a perfect match of power, speed, and performance for your applications •The Origin 200 server has high-performance processors, buses, and scalable I/O to keep up with your most complex application demands •The Origin 200 server was designed with embedded reliability, availability, and serviceability (RAS) so you can confidently trust your business to it •The Origin 200 server is easily expandable and upgradable—keeping pace with your demanding and changing business requirements •The Origin 200 server is a cost-effective business solution, both now and in the future Origin 200 is a sound server investment for your most important applications and is the gateway to the scalable Origin™ and SGI™ server product families. SGI offers an evolving portfolio of complete, pre-packaged solutions to enhance your productivity and success in areas such as Internet applications, media distribution, multiprotocol file serving, multitiered database management, and performance- intensive scientific or technical computing. -
SGI® Opengl Multipipe™ User's Guide
SGI® OpenGL Multipipe™ User’s Guide Version 2.3 007-4318-012 CONTRIBUTORS Written by Ken Jones and Jenn Byrnes Illustrated by Chrystie Danzer Production by Karen Jacobson Engineering contributions by Craig Dunwoody, Bill Feth, Alpana Kaulgud, Claude Knaus, Ravid Na’ali, Jeffrey Ungar, Christophe Winkler, Guy Zadicario, and Hansong Zhang COPYRIGHT © 2000–2003 Silicon Graphics, Inc. All rights reserved; provided portions may be copyright in third parties, as indicated elsewhere herein. No permission is granted to copy, distribute, or create derivative works from the contents of this electronic documentation in any manner, in whole or in part, without the prior written permission of Silicon Graphics, Inc. LIMITED RIGHTS LEGEND The electronic (software) version of this document was developed at private expense; if acquired under an agreement with the USA government or any contractor thereto, it is acquired as "commercial computer software" subject to the provisions of its applicable license agreement, as specified in (a) 48 CFR 12.212 of the FAR; or, if acquired for Department of Defense units, (b) 48 CFR 227-7202 of the DoD FAR Supplement; or sections succeeding thereto. Contractor/manufacturer is Silicon Graphics, Inc., 1600 Amphitheatre Pkwy 2E, Mountain View, CA 94043-1351. TRADEMARKS AND ATTRIBUTIONS Silicon Graphics, SGI, the SGI logo, InfiniteReality, IRIS, IRIX, Onyx, Onyx2, OpenGL, and Reality Center are registered trademarks and GL, InfinitePerformance, InfiniteReality2, IRIS GL, Octane2, Onyx4, Open Inventor, the OpenGL logo, OpenGL Multipipe, OpenGL Performer, Power Onyx, Tezro, and UltimateVision are trademarks of Silicon Graphics, Inc., in the United States and/or other countries worldwide. MIPS and R10000 are registered trademarks of MIPS Technologies, Inc. -
Computing @SERC Resources,Services and Policies
Computing @SERC Resources,Services and Policies R.Krishna Murthy SERC - An Introduction • A state-of-the-art Computing facility • Caters to the computing needs of education and research at the institute • Comprehensive range of systems to cater to a wide spectrum of computing requirements. • Excellent infrastructure supports uninterrupted computing - anywhere, all times. SERC - Facilities • Computing - – Powerful hardware with adequate resources – Excellent Systems and Application Software,tools and libraries • Printing, Plotting and Scanning services • Help-Desk - User Consultancy and Support • Library - Books, Manuals, Software, Distribution of Systems • SERC has 5 floors - Basement,Ground,First,Second and Third • Basement - Power and Airconditioning • Ground - Compute & File servers, Supercomputing Cluster • First floor - Common facilities for Course and Research - Windows,NT,Linux,Mac and other workstations Distribution of Systems - contd. • Second Floor – Access Stations for Research students • Third Floor – Access Stations for Course students • Both the floors have similar facilities Computing Systems Systems at SERC • ACCESS STATIONS *SUN ULTRA 20 Workstations – dual core Opteron 4GHz cpu, 1GB memory * IBM INTELLISTATION EPRO – Intel P4 2.4GHz cpu, 512 MB memory Both are Linux based systems OLDER Access stations * COMPAQ XP 10000 * SUN ULTRA 60 * HP C200 * SGI O2 * IBM POWER PC 43p Contd... FILE SERVERS 5TB SAN storage IBM RS/6000 43P 260 : 32 * 18GB Swappable SSA Disks. Contd.... • HIGH PERFORMANCE SERVERS * SHARED MEMORY MULTI PROCESSOR • IBM P-series 690 Regatta (32proc.,256 GB) • SGI ALTIX 3700 (32proc.,256GB) • SGI Altix 350 ( 16 proc.,16GB – 64GB) Contd... * IBM SP3. NH2 - 16 Processors WH2 - 4 Processors * Six COMPAQ ALPHA SERVER ES40 4 CPU’s per server with 667 MHz. -
Sgiconsole™ Hardware Connectivity Guide
SGIconsole™ Hardware Connectivity Guide Document Number 007-4340-001 Contributors Written by Francisco Razo Illustrated by Dan Young Production by Karen Jacobson Contributions by Jagdish Bhavsar, Michael T. Brown, Dick Brownell, Jason Chang, Steven Dean, Steve Ewing, Jim Friedl, Jim Grisham, Karen Johnson, Tony Kavadias, Paul Kinyon, Jenny Leung, Laraine MacKenzie, Philip Montalban, Rod Negus, Sonny Oh, Keith Rich, Laura Shepard, Paddy Sreenivasan, Rebecca Underwood, and Eric Zamost. COPYRIGHT © 2001 Silicon Graphics, Inc. All rights reserved; provided portions may be copyright in third parties, as indicated elsewhere herein. No permission is granted to copy, distribute, or create derivative works from the contents of this electronic documentation in any manner, in whole or in part, without the prior written permission of Silicon Graphics, Inc. LIMITED RIGHTS LEGEND The electronic (software) version of this document was developed at private expense; if acquired under an agreement with the USA government or any contractor thereto, it is acquired as "commercial computer software" subject to the provisions of its applicable license agreement, as specified in (a) 48 CFR 12.212 of the FAR; or, if acquired for Department of Defense units, (b) 48 CFR 227-7202 of the DoD FAR Supplement; or sections succeeding thereto. Contractor/manufacturer is Silicon Graphics, Inc., 1600 Amphitheatre Pkwy 2E, Mountain View, CA 94043-1351. TRADEMARKS AND ATTRIBUTIONS Indy, IRIS, IRIX, Onyx2, and Silicon Graphics are registered trademarks of Silicon Graphics, Inc. SGI, the SGI logo, IRISconsole, IRIS InSight, and SGIconsole are trademarks of Silicon Graphics, Inc. PostScript is a registered trademark of Adobe Systems, Inc. Linux is a registered trademark of Linus Torvalds. -
Fast Synchronization on Shared-Memory Multiprocessors: an Architectural Approach
Fast Synchronization on Shared-Memory Multiprocessors: An Architectural Approach Zhen Fang1, Lixin Zhang2, John B. Carter1, Liqun Cheng1, Michael Parker3 1 School of Computing University of Utah Salt Lake City, UT 84112, U.S.A. fzfang, retrac, [email protected] Telephone: 801.587.7910 Fax: 801.581.5843 2 IBM Austin Research Lab 3 Cray, Inc. 11400 Burnet Road, MS 904/6C019 1050 Lowater Road Austin, TX 78758, U.S.A. Chippewa Falls, WI 54729, U.S.A. [email protected] [email protected] Abstract Synchronization is a crucial operation in many parallel applications. Conventional synchronization mechanisms are failing to keep up with the increasing demand for efficient synchronization operations as systems grow larger and network latency increases. The contributions of this paper are threefold. First, we revisit some representative synchronization algorithms in light of recent architecture innovations and provide an example of how the simplifying assumptions made by typical analytical models of synchronization mechanisms can lead to significant performance estimate errors. Second, we present an architectural innovation called active memory that enables very fast atomic operations in a shared-memory multiprocessor. Third, we use execution-driven simulation to quantitatively compare the performance of a variety of synchronization mechanisms based on both existing hardware techniques and active memory operations. To the best of our knowledge, synchronization based on active memory outforms all existing spinlock and non-hardwired barrier implementations by a large margin. Keywords: distributed shared-memory, coherence protocol, synchronization, barrier, spinlock, memory controller This effort was supported by the National Security Agency (NSA), the Defense Advanced Research Projects Agency (DARPA) and Silicon Graphics Inc. -
Parallel Maximum-Likelihood Inversion for Estimating Wavenumber-Ordered Spectra in Emission Spectroscopy
Parallel Maximum-Likelihood Inversion for Estimating Wavenumber-Ordered Spectra in Emission Spectroscopy Hoda El-Sayed Marc Salit John Travis [email protected] [email protected] Judith Devaney William George [email protected] [email protected] National Institute of Standards and Technology Gaithersburg, Maryland USA Abstract distributes the signal-carried noise to the signal, preventing the masking of small spectral features by the signal-carried We introduce a parallelization of the maximum- noise from the large spectral features. The spectral esti- likelihood cosine transform. This transform consists of a mates obtained using maximum-likelihood inversion have computationally intensive iterative fitting process, but is another potentially useful property—a line-shape which is readily decomposed for parallel processing. The parallel burdened with a less distorting transform-function than the implementation is not only scalable, but has also brought sinc function of the Fourier transform. the execution time of this previously intractable problem to In this paper we present a parallel implementation of the feasible levels using contemporary and cost-efficient high- maximum-likelihood inversion method. It will be shown performance computers, including an SGI Origin 2000, an that the parallel implementation is not only scalable, but SGI Onyx, and a cluster of Intel-based PCs. has also brought the execution time of this problem to feasible levels using contemporary and cost-efficient high- Key words : parallel processing, emission spectroscopy, performance computers including Origin 2000, SGI Onyx, cosine transform, maximum-likelihood inversion, perfor- and PC clusters. By parallelizing this application, we were mance evaluation, DParLib, MPI. able to reduce the running time of the program to seconds rather than hours. -
İSTANBUL TECHNICAL UNIVERSITY INSTITUTE of SCIENCE and TECHNOLOGY Master Thesis by Göker Burak ÇETİN, Eng. Department
ĐSTANBUL TECHNICAL UNIVERSITY INSTITUTE OF SCIENCE AND TECHNOLOGY ITU-CSCRS GROUND RECEIVING STATION AUTOMATION & RENOVATION Master Thesis by Göker Burak ÇET ĐN, Eng. Department : Geodesy & Photogrammetry Engineering Programme: Geomatics Engineering SEPTEMBER 2007 ISTANBUL TECHNICAL UNIVERSITY INSTITUTE OF SCIENCE AND TECHNOLOGY ITU-CSCRS GROUND RECEIVING STATION AUTOMATION & RENOVATION Master Thesis by Göker Burak ÇET ĐN, Eng. 501031614 Date of submission : 4 September 2007 Date of defence examination: 3 September 2007 Supervisor (Chairman) : Prof. Dr. Filiz SUNAR Members of the Examining Committee : Prof. Dr. Derya MAKTAV (ITU) Assist. Prof. Dr. Berk ÜSTÜNDA Ğ (ITU) SEPTEMBER 2007 ĐSTANBUL TEKN ĐK ÜN ĐVERS ĐTES Đ FEN B ĐLĐMLER Đ ENST ĐTÜSÜ ĐTÜ-UHUZAM UYDU YER ĐSTASYONU OTOMASYON VE RENOVASYONU Master Tezi Müh. Göker Burak ÇET ĐN 501031614 Tezin Enstitüye Verildi ği Tarih : 4 Eylül 2007 Tezi Savunuldu ğu Tarih : 3 Eylül 2007 Tez Danı şmanı : Prof. Dr. Filiz SUNAR Di ğer Jüri Üyeleri : Prof. Dr. Derya MAKTAV ( Đ.T.Ü.) Yard. Doç. Dr. Berk ÜSTÜNDA Ğ ( Đ.T.Ü.) EYLÜL 2007 FOREWORD I have tried to solidate and provide this documentation related to satellite ground receiving station in a more methodical way after four years of labor and experience gained as a system administrator at ITU-CSCRS Ground Station. From the day I begin working at the ground station I observed some of the obstacles in the system. Now after 4 years with the aging hardware and software systems, it is now more obvious that the ground station needs a renovation project to carry on its operation. Therefore without constant renovation a ground station, could have never achieve enough service quality, because of the constant increase of demands of the remote sensing data market at the end user side. -
CXFSTM Administration Guide for SGI® Infinitestorage
CXFSTM Administration Guide for SGI® InfiniteStorage 007–4016–020 CONTRIBUTORS Written by Lori Johnson Illustrated by Chrystie Danzer Production by Karen Jacobson Engineering contributions to the book by Rich Altmaier, Neil Bannister, François Barbou des Places, Ken Beck, Felix Blyakher, Laurie Costello, Mark Cruciani, Dave Ellis, Brian Gaffey, Philippe Gregoire, Dean Jansa, Erik Jacobson, Dennis Kender, Chris Kirby, Ted Kline, Dan Knappe, Kent Koeninger, Linda Lait, Bob LaPreze, Steve Lord, Aaron Mantel, Troy McCorkell, LaNet Merrill, Terry Merth, Nate Pearlstein, Bryce Petty, Alain Renaud, John Relph, Elaine Robinson, Dean Roehrich, Eric Sandeen, Wesley Smith, Kerm Steffenhagen, Paddy Sreenivasan, Andy Tran, Rebecca Underwood, Connie Waring, Geoffrey Wehrman COPYRIGHT © 1999–2004 Silicon Graphics, Inc. All rights reserved; provided portions may be copyright in third parties, as indicated elsewhere herein. No permission is granted to copy, distribute, or create derivative works from the contents of this electronic documentation in any manner, in whole or in part, without the prior written permission of Silicon Graphics, Inc. LIMITED RIGHTS LEGEND The electronic (software) version of this document was developed at private expense; if acquired under an agreement with the USA government or any contractor thereto, it is acquired as "commercial computer software" subject to the provisions of its applicable license agreement, as specified in (a) 48 CFR 12.212 of the FAR; or, if acquired for Department of Defense units, (b) 48 CFR 227-7202 of -
Remanufactured SGI™ Origin™ 2000 High-Performance Server
Product Guide Remanufactured SGI™ Origin™ 2000 High-Performance Server Origin 2000 Rack Multirack System The rack-mounted Connect multiple racks as a shared- Origin 2000 system memory powerhouse or redeploy is the industry’s those assets into multiple, tightly most powerful and integrated departmental servers. flexible shared-memory Add IRIS FailSafe™ software to server. Easily expand provide failover between systems, the system by adding and Origin 2000 becomes the additional racks. ultimate high-availability server for critical computational and business solutions. Origin 2000 Deskside Revolutionary Server for Your Expanding Business SGI server family, can expand from dual-processor With its outstanding Today’s corporate environment is characterized by deskside systems to powerful 512-processor scalable price/performance, the a rapid growth in communications requirements servers without disruptive and expensive box-swaps deskside Origin 2000 and continual organizational change. A great deal common with other server solutions. System server is ideal for of energy is spent planning future system purchases resources can be deployed as tightly integrated evolving applications in an ongoing effort to keep system capabilities arrays of small Origin 2000 systems or as single requiring expansion capability. Seamless upgradability to larger ahead of application demands. Corporations require large shared-memory systems capable of meeting rack-based configurations supports scalable, flexible computer systems that can adapt the most demanding data processing and your growing requirements. to business change and satisfy growing server computing requirements. requirements. SGI offers you a proven, more adaptable way to purchase your data processing, SGI pioneered the development and delivery of computation, and communications capabilities. shared-memory parallel processing systems. -
IRIX® Device Driver Programmer's Guide
IRIX® Device Driver Programmer’s Guide Document Number 007-0911-210 CONTRIBUTORS Written by David Cortesi, John Raithel, Bill Tuthill, and Anita Manders UpdatedbyJulieBoneyandStevenLevine Illustrated by Dany Galgani, Cheri Brown, and Chrystie Danzer Production by Karen Jacobson COPYRIGHT © 1998-2003 Silicon Graphics, Inc. All rights reserved; provided portions may be copyright in third parties, as indicated elsewhere herein. No permission is granted to copy, distribute, or create derivative works from the contents of this electronic documentation in any manner, in whole or in part, without the prior written permission of Silicon Graphics, Inc. LIMITED RIGHTS LEGEND The electronic (software) version of this document was developed at private expense; if acquired under an agreement with the USA government or any contractor thereto, it is acquired as "commercial computer software" subject to the provisions of its applicable license agreement, as specified in (a) 48 CFR 12.212 of the FAR; or, if acquired for Department of Defense units, (b) 48 CFR 227-7202 of the DoD FAR Supplement; or sections succeeding thereto. Contractor/manufacturer is Silicon Graphics, Inc., 1600 Amphitheatre Pkwy 2E, Mountain View, CA 94043-1351. TRADEMARKS AND ATTRIBUTIONS Silicon Graphics, SGI, the SGI logo, Challenge, Indigo, IRIS, IRIX, O2, Octane, Onyx, Onyx2, and Origin are registered trademarks and Indigo2,Indigo2 Maximum Impact, IRIS InSight, Power Challenge, Power Channel, Power Indigo2,PowerOnyx,andREACT/proare trademarks of Silicon Graphics, Inc., in the United States and/or other countries worldwide. Indy is a registered trademark, used under license in the United States and owned by Silicon Graphics,Inc.inothercountriesworldwide. IBM is a trademark of International Business Machines Corporation.