SGI™ Origin 3000 Series Technical Configuration Owner's Guide
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Paper in PDF Format
Managing System Resources on ‘Teras’ Experiences in Managing System Resources on a Large SGI Origin 3800 Cluster The paper to the presentation given at the CUG SUMMIT 2002 held at the University of Manchester. Manchester (UK), May 20-24 2002. http://www.sara.nl Mark van de Sanden [email protected] Huub Stoffers [email protected] Introduction Outline of the Paper This is a paper about our experiences with managing system resources on ‘Teras’. Since November 2000 Teras, a large SGI Origin 3800 cluster, is the national supercomputer for the Dutch academic community. Teras is located at and maintained by SARA High Performance Computing (HPC). The outline of this presentation is as follows: Very briefly, we will tell you what SARA is, and what we presently do, particularly in high performance computing, besides maintaining Teras. Subsequently, to give you an idea, on the one hand which resources are available on Teras, and on the other hand what resources are demanded, we will give overviews of three particular aspects of the Teras: 1) Some data on the hardware – number of CPUs, memory, etc. - and the cluster configuration, i.e. description of the hosts that constitute the cluster. 2) Identification of the key software components that are used on the system; both, system software components, as well as software toolkits available to users to create their (parallel) applications. 3) A characterization of the ‘job mix’ that runs on the Teras cluster. Thus equipped with a clearer idea of both ‘supply’ and ‘demand’ of resources on Teras, we then state the resource allocation policy goals that we pursue in this context, and we review what resource management possibilities we have at hand and what they can accomplish for us, in principle. -
SGI® Octane III®
Making ® ® Supercomputing SGI Octane III Personal™ KEY FEATURES Supercomputing Gets Personal Octane III takes high-performance computing out Office Ready of the data center and puts it at the deskside. It A pedestal, one foot by two combines the immense power and performance broad HPC application support and arrives ready foot form factor and capabilities of a high-performance cluster with the for immediate integration for a smooth out-of-the- quiet operation portability and usability of a workstation to enable box experience. a new era of personal innovation in strategic science, research, development and visualization. Octane III allows a wide variety of single and High Performance dual-socket node choices and a wide selection of Up to 120 high- In contrast with standard dual-processor performance, storage, integrated networking, and performance cores and workstations with only eight cores and moderate graphics and compute GPU options. The system nearly 2TB of memory memory capacity, the superior design of is available as an up to ten node deskside cluster Broad HPC Application Octane III permits up to 120 high-performance configuration or dual-node graphics workstation Support cores and nearly 2TB of memory. Octane III configurations. Accelerated time-to-results significantly accelerates time-to-results for over with support for over 50 50 HPC applications and supports the latest Supported operating systems include: ® ® ® HPC applications Intel processors to capitalize on greater levels SUSE Linux Enterprise Server and Red Hat of performance, flexibility and scalability. Pre- Enterprise Linux. All configurations are available configured with system software, cluster set up is with pre-loaded SGI® Performance Suite system Ease of Use a breeze. -
Overview and History Nas Overview
zjjvojvovo OVERVIEWNAS OVERVIEW AND HISTORY a Look at NAS’ 25 Years of Innovation and a Glimpse at the Future In the mid-1970s, a group of Ames aerospace researchers began to study a highly innovative concept: NASA could transform U.S. aerospace R&D from the costly and cghghmgm time-consuming wind tunnel-based process to simulation-centric design and engineer- ing by executing emerging computational fluid dynamics (CFD) models on supercom- puters at least 1,000 times more powerful than those commercially available at the time. In 1976, Ames Center Director Dr. Hans Mark tasked a group led by Dr. F. Ronald Bailey to explore this concept, leading to formation of the Numerical Aerodynamic Simulator (NAS) Projects Office in 1979. At the same time, a user interface group was formed consisting of CFD leaders from industry, government, and academia to help guide requirements for the NAS concept gmfgfgmfand provide feedback on evolving computer feasibility studies. At the conclusion of these activities in 1982, NASA management changed the NAS approach from a focus on purchasing a specially developed supercomputer to an on-going Numerical Aerody- namic Simulation Program to provide leading-edge computational capabilities based on an innovative network-centric environment. The NAS Program plan for implementing this new approach was signed on February 8, 1983. Grand Opening As the NAS Program got underway, a projects office to a full-fledged division its first supercomputers were installed at Ames. In January 1987, NAS staff and in Ames’ Central Computing Facility, equipment were relocated to the new starting with a Cray X-MP-12 in 1984. -
OCTANE Technical Report
OCTANE Technical Report Silicon Graphics, Inc. ANY DUPLICATION, MODIFICATION, DISTRIBUTION, PUBLIC PERFORMANCE, OR PUBLIC DISPLAY OF THIS DOCUMENT WITHOUT THE EXPRESS WRITTEN CONSENT OF SILICON GRAPHICS, INC. IS STRICTLY PROHIBITED. THE RECEIPT OR POSSESSION OF THIS DOCUMENT DOES NOT CONVEY ANY RIGHTS TO REPRODUCE, DISCLOSE OR DISTRIBUTE ITS CONTENTS, OR TO MANUFACTURE, USE, OR SELL ANYTHING THAT IT MAY DESCRIBE, IN WHOLE OR IN PART. Cop VED Table of Contents Section 1 Introduction .....................................................................................................1-1 1.1 Manufacturing.............................................................................................................. 1-3 1.1.1 Industrial Design .....................................................................................1-3 1.1.2 CAD/CAM Solid Modeling ....................................................................1-4 1.1.3 Analysis...................................................................................................1-5 1.1.4 Digital Prototyping..................................................................................1-6 1.2 Entertainment............................................................................................................... 1-8 1.2.1 3D Animation..........................................................................................1-8 1.2.2 Film/Video/Audio ...................................................................................1-9 1.2.3 Publishing..............................................................................................1-10 -
Interactive Rendering with Coherent Ray Tracing
EUROGRAPHICS 2001 / A. Chalmers and T.-M. Rhyne Volume 20 (2001), Number 3 (Guest Editors) Interactive Rendering with Coherent Ray Tracing Ingo Wald, Philipp Slusallek, Carsten Benthin, and Markus Wagner Computer Graphics Group, Saarland University Abstract For almost two decades researchers have argued that ray tracing will eventually become faster than the rasteri- zation technique that completely dominates todays graphics hardware. However, this has not happened yet. Ray tracing is still exclusively being used for off-line rendering of photorealistic images and it is commonly believed that ray tracing is simply too costly to ever challenge rasterization-based algorithms for interactive use. However, there is hardly any scientific analysis that supports either point of view. In particular there is no evidence of where the crossover point might be, at which ray tracing would eventually become faster, or if such a point does exist at all. This paper provides several contributions to this discussion: We first present a highly optimized implementation of a ray tracer that improves performance by more than an order of magnitude compared to currently available ray tracers. The new algorithm makes better use of computational resources such as caches and SIMD instructions and better exploits image and object space coherence. Secondly, we show that this software implementation can challenge and even outperform high-end graphics hardware in interactive rendering performance for complex environments. We also provide an brief overview of the benefits of ray tracing over rasterization algorithms and point out the potential of interactive ray tracing both in hardware and software. 1. Introduction Ray tracing is famous for its ability to generate high-quality images but is also well-known for long rendering times due to its high computational cost. -
SGI® Opengl Multipipe™ User's Guide
SGI® OpenGL Multipipe™ User’s Guide Version 2.3 007-4318-012 CONTRIBUTORS Written by Ken Jones and Jenn Byrnes Illustrated by Chrystie Danzer Production by Karen Jacobson Engineering contributions by Craig Dunwoody, Bill Feth, Alpana Kaulgud, Claude Knaus, Ravid Na’ali, Jeffrey Ungar, Christophe Winkler, Guy Zadicario, and Hansong Zhang COPYRIGHT © 2000–2003 Silicon Graphics, Inc. All rights reserved; provided portions may be copyright in third parties, as indicated elsewhere herein. No permission is granted to copy, distribute, or create derivative works from the contents of this electronic documentation in any manner, in whole or in part, without the prior written permission of Silicon Graphics, Inc. LIMITED RIGHTS LEGEND The electronic (software) version of this document was developed at private expense; if acquired under an agreement with the USA government or any contractor thereto, it is acquired as "commercial computer software" subject to the provisions of its applicable license agreement, as specified in (a) 48 CFR 12.212 of the FAR; or, if acquired for Department of Defense units, (b) 48 CFR 227-7202 of the DoD FAR Supplement; or sections succeeding thereto. Contractor/manufacturer is Silicon Graphics, Inc., 1600 Amphitheatre Pkwy 2E, Mountain View, CA 94043-1351. TRADEMARKS AND ATTRIBUTIONS Silicon Graphics, SGI, the SGI logo, InfiniteReality, IRIS, IRIX, Onyx, Onyx2, OpenGL, and Reality Center are registered trademarks and GL, InfinitePerformance, InfiniteReality2, IRIS GL, Octane2, Onyx4, Open Inventor, the OpenGL logo, OpenGL Multipipe, OpenGL Performer, Power Onyx, Tezro, and UltimateVision are trademarks of Silicon Graphics, Inc., in the United States and/or other countries worldwide. MIPS and R10000 are registered trademarks of MIPS Technologies, Inc. -
Computing @SERC Resources,Services and Policies
Computing @SERC Resources,Services and Policies R.Krishna Murthy SERC - An Introduction • A state-of-the-art Computing facility • Caters to the computing needs of education and research at the institute • Comprehensive range of systems to cater to a wide spectrum of computing requirements. • Excellent infrastructure supports uninterrupted computing - anywhere, all times. SERC - Facilities • Computing - – Powerful hardware with adequate resources – Excellent Systems and Application Software,tools and libraries • Printing, Plotting and Scanning services • Help-Desk - User Consultancy and Support • Library - Books, Manuals, Software, Distribution of Systems • SERC has 5 floors - Basement,Ground,First,Second and Third • Basement - Power and Airconditioning • Ground - Compute & File servers, Supercomputing Cluster • First floor - Common facilities for Course and Research - Windows,NT,Linux,Mac and other workstations Distribution of Systems - contd. • Second Floor – Access Stations for Research students • Third Floor – Access Stations for Course students • Both the floors have similar facilities Computing Systems Systems at SERC • ACCESS STATIONS *SUN ULTRA 20 Workstations – dual core Opteron 4GHz cpu, 1GB memory * IBM INTELLISTATION EPRO – Intel P4 2.4GHz cpu, 512 MB memory Both are Linux based systems OLDER Access stations * COMPAQ XP 10000 * SUN ULTRA 60 * HP C200 * SGI O2 * IBM POWER PC 43p Contd... FILE SERVERS 5TB SAN storage IBM RS/6000 43P 260 : 32 * 18GB Swappable SSA Disks. Contd.... • HIGH PERFORMANCE SERVERS * SHARED MEMORY MULTI PROCESSOR • IBM P-series 690 Regatta (32proc.,256 GB) • SGI ALTIX 3700 (32proc.,256GB) • SGI Altix 350 ( 16 proc.,16GB – 64GB) Contd... * IBM SP3. NH2 - 16 Processors WH2 - 4 Processors * Six COMPAQ ALPHA SERVER ES40 4 CPU’s per server with 667 MHz. -
OCTANE® Workstation Owner's Guide
OCTANE® Workstation Owner’s Guide Document Number 007-3435-003 CONTRIBUTORS Written by Charmaine Moyer Production by Linda Rae Sande Illustrated by Kwong Liew Engineering contributions by Jim Bergman, Brian Bolich, Bob Cook, Mark Glusker, John Hahn, Steve Manzi, Ted Marsh, Donna McMaster, Jim Pagura, Michael Poimboeuf, Brad Reger, Jose Reinoso, Bob Sanders, Chris Wheaton, Michael Wright, and many others on the OCTANE engineering and business team. St. Peter’s Basilica image courtesy of ENEL SpA and InfoByte SpA. Disk Thrower image courtesy of Xavier Berenguer, Animatica. © 1997 - 1999, Silicon Graphics, Inc.— All Rights Reserved The contents of this document may not be copied or duplicated in any form, in whole or in part, without the prior written permission of Silicon Graphics, Inc. LIMITED AND RESTRICTED RIGHTS LEGEND Use, duplication, or disclosure by the Government is subject to restrictionsas set forth in the Rights in Data clause at FAR 52.227-14 and/or in similar orsuccessor clauses in the FAR, or in the DOD, DOE or NASA FAR Supplements.Unpublished rights reserved under the Copyright Laws of the United States.Contractor/manufacturer is Silicon Graphics, Inc., 2011 N. Shoreline Blvd., Mountain View, CA 94043-1389. Silicon Graphics, IRIS, IRIX, and OCTANE are registered trademarks and the Silicon Graphics logo, IRIX Interactive Desktop, Power Fortran Accelerator, IRIS InSight, and Stereoview are trademarks of Silicon Graphics, Inc. ADAT is a registered trademark of Alesis Corporation. Centronics is a registered trademark of Centronics Data Computer Corporation. Envi-ro-tech is a trademark of TECHSPRAY. Macintosh is a registered trademark of Apple Computer, Inc. -
Fast Synchronization on Shared-Memory Multiprocessors: an Architectural Approach
Fast Synchronization on Shared-Memory Multiprocessors: An Architectural Approach Zhen Fang1, Lixin Zhang2, John B. Carter1, Liqun Cheng1, Michael Parker3 1 School of Computing University of Utah Salt Lake City, UT 84112, U.S.A. fzfang, retrac, [email protected] Telephone: 801.587.7910 Fax: 801.581.5843 2 IBM Austin Research Lab 3 Cray, Inc. 11400 Burnet Road, MS 904/6C019 1050 Lowater Road Austin, TX 78758, U.S.A. Chippewa Falls, WI 54729, U.S.A. [email protected] [email protected] Abstract Synchronization is a crucial operation in many parallel applications. Conventional synchronization mechanisms are failing to keep up with the increasing demand for efficient synchronization operations as systems grow larger and network latency increases. The contributions of this paper are threefold. First, we revisit some representative synchronization algorithms in light of recent architecture innovations and provide an example of how the simplifying assumptions made by typical analytical models of synchronization mechanisms can lead to significant performance estimate errors. Second, we present an architectural innovation called active memory that enables very fast atomic operations in a shared-memory multiprocessor. Third, we use execution-driven simulation to quantitatively compare the performance of a variety of synchronization mechanisms based on both existing hardware techniques and active memory operations. To the best of our knowledge, synchronization based on active memory outforms all existing spinlock and non-hardwired barrier implementations by a large margin. Keywords: distributed shared-memory, coherence protocol, synchronization, barrier, spinlock, memory controller This effort was supported by the National Security Agency (NSA), the Defense Advanced Research Projects Agency (DARPA) and Silicon Graphics Inc. -
Parallel Maximum-Likelihood Inversion for Estimating Wavenumber-Ordered Spectra in Emission Spectroscopy
Parallel Maximum-Likelihood Inversion for Estimating Wavenumber-Ordered Spectra in Emission Spectroscopy Hoda El-Sayed Marc Salit John Travis [email protected] [email protected] Judith Devaney William George [email protected] [email protected] National Institute of Standards and Technology Gaithersburg, Maryland USA Abstract distributes the signal-carried noise to the signal, preventing the masking of small spectral features by the signal-carried We introduce a parallelization of the maximum- noise from the large spectral features. The spectral esti- likelihood cosine transform. This transform consists of a mates obtained using maximum-likelihood inversion have computationally intensive iterative fitting process, but is another potentially useful property—a line-shape which is readily decomposed for parallel processing. The parallel burdened with a less distorting transform-function than the implementation is not only scalable, but has also brought sinc function of the Fourier transform. the execution time of this previously intractable problem to In this paper we present a parallel implementation of the feasible levels using contemporary and cost-efficient high- maximum-likelihood inversion method. It will be shown performance computers, including an SGI Origin 2000, an that the parallel implementation is not only scalable, but SGI Onyx, and a cluster of Intel-based PCs. has also brought the execution time of this problem to feasible levels using contemporary and cost-efficient high- Key words : parallel processing, emission spectroscopy, performance computers including Origin 2000, SGI Onyx, cosine transform, maximum-likelihood inversion, perfor- and PC clusters. By parallelizing this application, we were mance evaluation, DParLib, MPI. able to reduce the running time of the program to seconds rather than hours. -
AVS on UNIX WORKSTATIONS INSTALLATION/ RELEASE NOTES
_________ ____ AVS on UNIX WORKSTATIONS INSTALLATION/ RELEASE NOTES ____________ Release 5.5 Final (50.86 / 50.88) November, 1999 Advanced Visual Systems Inc.________ Part Number: 330-0120-02 Rev L NOTICE This document, and the software and other products described or referenced in it, are con®dential and proprietary products of Advanced Visual Systems Inc. or its licensors. They are provided under, and are subject to, the terms and conditions of a written license agreement between Advanced Visual Systems and its customer, and may not be transferred, disclosed or otherwise provided to third parties, unless oth- erwise permitted by that agreement. NO REPRESENTATION OR OTHER AFFIRMATION OF FACT CONTAINED IN THIS DOCUMENT, INCLUDING WITHOUT LIMITATION STATEMENTS REGARDING CAPACITY, PERFORMANCE, OR SUI- TABILITY FOR USE OF SOFTWARE DESCRIBED HEREIN, SHALL BE DEEMED TO BE A WARRANTY BY ADVANCED VISUAL SYSTEMS FOR ANY PURPOSE OR GIVE RISE TO ANY LIABILITY OF ADVANCED VISUAL SYSTEMS WHATSOEVER. ADVANCED VISUAL SYSTEMS MAKES NO WAR- RANTY OF ANY KIND IN OR WITH REGARD TO THIS DOCUMENT, INCLUDING BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PUR- POSE. ADVANCED VISUAL SYSTEMS SHALL NOT BE RESPONSIBLE FOR ANY ERRORS THAT MAY APPEAR IN THIS DOCUMENT AND SHALL NOT BE LIABLE FOR ANY DAMAGES, INCLUDING WITHOUT LIMI- TATION INCIDENTAL, INDIRECT, SPECIAL OR CONSEQUENTIAL DAMAGES, ARISING OUT OF OR RELATED TO THIS DOCUMENT OR THE INFORMATION CONTAINED IN IT, EVEN IF ADVANCED VISUAL SYSTEMS HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH DAMAGES. The speci®cations and other information contained in this document for some purposes may not be com- plete, current or correct, and are subject to change without notice. -
Running.1997.2.Pdf
SEMI-ANNUAL REPORT NASA CONTRACT NAS5-31368 FOR MODIS TEAM MEMBER STEVEN W. RUNNING ASSOC. TEAM MEMBER RAMAKRISHNA R. NEMANI SOFTWARE ENGINEER JOSEPH GLASSY July 15, 1997 PRE-LAUNCH TASKS PROPOSED IN OUR CONTRACT OF DECEMBER 1991 We propose, during the pre-EOS phase to: (1) develop, with other MODIS Team Members, a means of discriminating different major biome types with NDVI and other AVHRR-based data. (2) develop a simple ecosystem process model for each of these biomes, BIOME-BGC (3) relate the seasonal trend of weekly composite NDVI to vegetation phenology and temperature limits to develop a satellite defined growing season for vegetation; and (4) define physiologically based energy to mass conversion factors for carbon and water for each biome. Our final core at-launch product will be simplified, completely satellite driven biome specific models for net primary production. We will build these biome specific satellite driven algorithms using a family of simple ecosystem process models as calibration models, collectively called BIOME-BGC, and establish coordination with an existing network of ecological study sites in order to test and validate these products. Field datasets will then be available for both BIOME-BGC development and testing, use for algorithm developments of other MODIS Team Members, and ultimately be our first test point for MODIS land vegetation products upon launch. We will use field sites from the National Science Foundation Long-Term Ecological Research network, and develop Glacier National Park as a major site for intensive validation. OBJECTIVES: We have defined the following near-term objectives for our MODIS contract based on the long term objectives stated above: -Organization of an EOS ground monitoring network with collaborating U.S.