The Development and Performance of a Message-Passing Version of the PAGOSA Shock-Wave Physics Code

Total Page:16

File Type:pdf, Size:1020Kb

The Development and Performance of a Message-Passing Version of the PAGOSA Shock-Wave Physics Code SAND97-2551 Distribution Unlimited Release Category UC-705 Printed October 1997 The Development and Performance of a Message-Passing Version of the PAGOSA Shock-Wave Physics Code David R. Gardner Parallel Computational Sciences Department Courtenay T. Vaughan Parallel Computing Science Department P.O. Box 5800 Sandia National Laboratories Albuquerque, NM 87185-111 Prepared for The Computational Mechanics and Material Modeling Technology Coordination Group of the Joint DoD/DOE Munitions Technology Development Program Abstract A message-passing version of the PAGOSA shock-wave physics code has been developed at Sandia National Laboratories for multiple-instruction, multiple-data stream (MIMD) computers. PAGOSA is an explicit, Eulerian code for modeling the three-dimensional, high-speed hydrodynamic flow of fluids and the dynamic deformation of solids under high rates of strain. It was originally developed at Los Alamos National Laboratory for the sin- gle-instruction, multiple-data (SIMD) Connection Machine parallel com- puters. The performance of Sandia’s message-passing version of PAGOSA has been measured on two MIMD machines, the nCUBE 2 and the Intel Paragon XP/S. No special efforts were made to optimize the code for ei- ther machine. The measured scaled speedup (computational time for a sin- gle computational node divided by the computational time per node for fixed computational load) and grind time (computational time per cell per time step) show that the MIMD PAGOSA code scales linearly with the number of computational nodes used on a variety of problems, including the simulation of shaped-charge jets perforating an oil well casing. Scaled parallel efficiencies for MIMD PAGOSA are greater than 0.70 when the available memory per node is filled (or nearly filled) on hundreds to a thou- sand or more computational nodes on these two machines, indicating that the code scales very well. Thus good parallel performance can be achieved for complex and realistic applications when they are first implemented on MIMD parallel computers. 1 Acknowledgments The PAGOSA code was developed at Los Alamos National Laboratory under the direction of Dr. J. W. Hopson. We thank Dr. Hopson for making the data-parallel version of PAGOSA available to us, and we thank Dr. D. B. Kothe of Los Alamos National Laborato- ry for providing technical advice concerning its structure. We thank the Advanced Com- puting Laboratory of Los Alamos National Laboratory, Los Alamos, NM 87545, for providing computing resources necessary to complete this work. This work was supported under the Joint DoD/DOE Munitions Technology Development Program, and sponsored by the Office of Munitions of the Secretary of Defense. We thank T. Mack Stallcup of Intel Corporation for providing technical information about the Paragon computer. We thank Martin W. Lewitt, formerly of nCUBE Corporation, for technical information about the nCUBE 2 computer. UNIX® is a trademark of AT&T. CM-2, and CM-5 are trademarks of Thinking Machines Corporation. iPSC®, i860, and Paragon™ are trademarks of Intel Corporation. This work was performed at Sandia National Laboratories supported by the U.S. Depart- ment of Energy under contract number DE-AC04-94AL85000. We thank the Massively Parallel Computing Research Laboratory at Sandia for providing computing resources necessary to complete this work. 2 Table of Contents Abstract................................................................................................................................1 Acknowledgments................................................................................................................2 Introduction..........................................................................................................................9 Issues in Parallel Computing .............................................................................................11 Parallel Code Performance Measurements ........................................................................13 Development of the MIMD PAGOSA Code.......................................................................18 Features of PAGOSA 5.5 ........................................................................................18 Development of MIMD PAGOSA 5.5 from SIMD PAGOSA 5.5...........................19 Features of MIMD PAGOSA 5.5............................................................................22 The Test Simulations .........................................................................................................22 The Finned Projectile Simulations.........................................................................22 The Explosive Welding Simulation.......................................................................23 The Oil-Well Perforation Simulation.....................................................................24 The Test Conditions...........................................................................................................27 The Performance of MIMD PAGOSA ..............................................................................28 Message-Passing Performance on the nCUBE 2 ...................................................29 Message-Passing Performance on the Intel Paragon .........................................36 Performance of MIMD PAGOSA on the nCUBE 2 and the Intel Paragon Message-Passing Computers .............................................................................................46 Summary and Conclusions ................................................................................................52 References..........................................................................................................................54 Appendices.........................................................................................................................56 A MIMD PAGOSA 5.5 Input Guide.....................................................................56 B Test Problem Input Sets...................................................................................90 B1 Input Set for the Finned Projectile Problem with the Hydrodynamic Constitutive Model, fp1 .............................................................................90 3 B2 Input Set for the Finned Projectile Problem with the Elastic, Perfectly Plastic Constitutive Model, fp2 .................................................................94 B3 Input Set for the Explosive Welding Problem, ew ....................................98 B4 Input Set for the Oil-Well Perforation Problem, owp..............................102 4 List of Figures 1 Illustration of the speedup surface, the fixed-size speedup curve, and the scaled speedup curve.................................................................................................15 2 The fixed-size speedup and scaled speedup curves projected on the P-S plane ...................................................................................................................17 3 Illustration of the decomposition of the global computational domain into subdomains, in two spatial dimensions.....................................................................20 4 Program fragments illustrating the translation from CM Fortran to Fortran 77.......20 5 Illustration of shift_left local inter-node communication between two nodes for two-dimensional subdomains for a message-passing code.......................................21 6 Simulation of a finned tungsten projectile obliquely impacting a stainless steel plate (hydrodynamic constitutive model) .................................................................23 7 Simulation of a finned tungsten projectile obliquely impacting a stainless steel plate (elastic, perfectly plastic constitutive model)...................................................23 8 Simulation of the explosive welding of a copper tube to a stainless steel plate. ......24 9 Simulation of the perforation of a steel oil-well casing by a shaped charge jet. ......26 10 Scaled speedup and grind time for the fp1 simulation on the nCUBE 2. .................33 11 Scaled speedup and grind time for the fp2 simulation on the nCUBE 2. .................33 12 Scaled speedup and grind time for the ew simulation on the nCUBE 2...................34 13 Scaled speedup and grind time for the owp simulation on the nCUBE 2.................34 14 Scaled parallel efficiency as a function of the number of nodes on the nCUBE 2...35 15 Scaled speedup and grind time for the fp1 simulation on the Paragon.................43 16 Scaled speedup and grind time for the fp2 simulation on the Paragon.................43 17 Scaled speedup and grind time for the ew simulation on the Paragon..................44 18 Scaled speedup and grind time for the owp simulation on the Paragon. ..............44 19 Scaled parallel efficiency on the Paragon .............................................................45 20 Effect of subdomain size on the scaled parallel efficiency on the Paragon for the ew and owp simulations......................................................................................45 5 Intentionally Blank Page 6 List of Tables 1 Compiler Versions and Options for Compiling MIMD PAGOSA..............................27 2 Grind Time Repeatability ............................................................................................27 3 Performance of MIMD PAGOSA on the nCUBE 2 for fp1..........................................31 4 Performance of MIMD PAGOSA on the nCUBE 2 for fp2..........................................31
Recommended publications
  • Ada Departmental Supercomputer Shared Memory GPU Cluster
    Ada Departmental Supercomputer Shared Memory GPU Cluster The Ada Departmental Supercomputer is designed to provide System Specifications near top 500 class supercomputing capabilities at your office Processors: Head Node: 2 AMD EPYC 7702 Processors or lab. (64 core-2.0/3.3 GHz) Compute Nodes: 1 AMD EPYC 7702P Proces- Ada is a hybrid supercomputer consisting of a large memory sor (64 core-2.2/3.2 GHz), 8 AMD Radeon head node and 2 to 5 compute nodes, each with eight AMD Instinct MI50 GPUs Radeon Instinct MI50 GPUs. With 5 compute nodes Ada con- Global Memory: 2TB or 4TB 3200 MHz DDR4 tains 448 AMD EPYC processor cores, 40 MI50 GPUs and 2 or 4 TB of globally shared memory. The compute nodes are Compute Node 128 GB 3200 MHz DDR4 (each) Memory: connected to the head node with 200 Gb/s Mellanox Infini- band. The Ada departmental supercomputer can be config- Storage: 1TB on-board M.2 OS SSD ured to deliver 1060 TFLOPS of FP16, 532 TFLOPS of FP32 12x 3.5" SATA/SAS hot-swap and 264 TFLOPS of FP64 GPU floating point performance SSD/HDD bays (head node) Additional 8x 2.5” SSD hot-swap bays on each capable of operating on large computational models. compute node Ada is a true symmetric multi-processing (SMP) computer Interconnect: ConnectX-6 VPI 200 Gb/s InfiniBand Dual Port with a large shared memory and a single operating system PCIe Gen 4 Host Bus Adapters user interface based on Centos 8 Linux. It provides a 1TB (No InfiniBand switch is needed) globally shared fast file system, and a large disk storage ar- I/O: 2x 1 Gb/s LAN ports ray.
    [Show full text]
  • Sprite File System There Are Three Important Aspects of the Sprite ®Le System: the Scale of the System, Location-Transparency, and Distributed State
    Naming, State Management, and User-Level Extensions in the Sprite Distributed File System Copyright 1990 Brent Ballinger Welch CHAPTER 1 Introduction ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ This dissertation concerns network computing environments. Advances in network and microprocessor technology have caused a shift from stand-alone timesharing systems to networks of powerful personal computers. Operating systems designed for stand-alone timesharing hosts do not adapt easily to a distributed environment. Resources like disk storage, printers, and tape drives are not concentrated at a single point. Instead, they are scattered around the network under the control of different hosts. New operating system mechanisms are needed to handle this sort of distribution so that users and application programs need not worry about the distributed nature of the underlying system. This dissertation explores the approach of centering a distributed computing environment around a shared network ®le system. The ®le system is chosen as a starting point because it is a heavily used service in stand-alone systems, and the read/write para- digm of the ®le system is a familiar one that can be applied to many system resources. The ®le system described in this dissertation provides a distributed name space for sys- tem resources, and it provides remote access facilities so all resources are available throughout the network. Resources accessible via the ®le system include disk storage, other types of peripheral devices, and user-implemented service applications. The result- ing system is one where resources are named and accessed via the shared ®le system, and the underlying distribution of the system among a collection of hosts is not important to users.
    [Show full text]
  • HDF-EOS Interface Based on HDF5, Volume 2: Function Reference Guide Was Prepared Under the EOSDIS Evolution and Development-2 Contract (NNG15HZ39C)
    This page intentionally left blank. Preface This document is a Users Guide for HDF-EOS (Hierarchical Data Format - Earth Observing System) library tools. The version described in this document is HDF-EOS Version 5.1.16. The software is based on HDF5, a new version of HDF provided by by The HDF Group. HDF5 is a complete rewrite of the earlier HDF4 version, containing a different data model and user interface. HDF-EOS V5.1.16 incorporates HDF5, and keeps the familier HDF4-based interface. There are a few exceptions and these exceptions are described in this document. Note that the major functional difference is that Version 5.1.16 of the HDF-EOS library is a thread-safe. HDF is the scientific data format standard selected by NASA as the baseline standard for EOS. This Users Guide accompanies Version 5.1.16 software, which is available to the user community on the EDHS1 server. This library is aimed at EOS data producers and consumers, who will develop their data into increasingly higher order products. These products range from calibrated Level 1 to Level 4 model data. The primary use of the HDF-EOS library will be to create structures for associating geolocation data with their associated science data. This association is specified by producers through use of the supplied library. Most EOS data products which have been identified, fall into categories of Point, Grid, Swath or Zonal Average structures, the latter two of which are implemented in the current version of the library. Services based on geolocation information will be built on HDF-EOS structures.
    [Show full text]
  • Cielo Computational Environment Usage Model
    Cielo Computational Environment Usage Model With Mappings to ACE Requirements for the General Availability User Environment Capabilities Release Version 1.1 Version 1.0: Bob Tomlinson, John Cerutti, Robert A. Ballance (Eds.) Version 1.1: Manuel Vigil, Jeffrey Johnson, Karen Haskell, Robert A. Ballance (Eds.) Prepared by the Alliance for Computing at Extreme Scale (ACES), a partnership of Los Alamos National Laboratory and Sandia National Laboratories. Approved for public release, unlimited dissemination LA-UR-12-24015 July 2012 Los Alamos National Laboratory Sandia National Laboratories Disclaimer Unless otherwise indicated, this information has been authored by an employee or employees of the Los Alamos National Security, LLC. (LANS), operator of the Los Alamos National Laboratory under Contract No. DE-AC52-06NA25396 with the U.S. Department of Energy. The U.S. Government has rights to use, reproduce, and distribute this information. The public may copy and use this information without charge, provided that this Notice and any statement of authorship are reproduced on all copies. Neither the Government nor LANS makes any warranty, express or implied, or assumes any liability or responsibility for the use of this information. Bob Tomlinson – Los Alamos National Laboratory John H. Cerutti – Los Alamos National Laboratory Robert A. Ballance – Sandia National Laboratories Karen H. Haskell – Sandia National Laboratories (Editors) Cray, LibSci, and PathScale are federally registered trademarks. Cray Apprentice2, Cray Apprentice2 Desktop, Cray C++ Compiling System, Cray Fortran Compiler, Cray Linux Environment, Cray SHMEM, Cray XE, Cray XE6, Cray XT, Cray XTm, Cray XT3, Cray XT4, Cray XT5, Cray XT5h, Cray XT5m, Cray XT6, Cray XT6m, CrayDoc, CrayPort, CRInform, Gemini, Libsci and UNICOS/lc are trademarks of Cray Inc.
    [Show full text]
  • EOS: a Project to Investigate the Design and Construction of Real-Time Distributed Embedded Operating Systems
    c EOS: A Project to Investigate the Design and Construction of Real-Time Distributed Embedded Operating Systems. * (hASA-CR-18G971) EOS: A PbCJECZ 10 187-26577 INVESTIGATE TEE CESIGI AND CCES!I&CCIXOti OF GEBL-1IIBE DISZEIEOTEO EWBEECIC CEERATIN6 SPSTEI!!S Bid-Year lieport, 1QE7 (Illinois Unclas Gniv.) 246 p Avail: AlIS BC All/!!P A01 63/62 00362E8 Principal Investigator: R. H. Campbell. Research Assistants: Ray B. Essick, Gary Johnston, Kevin Kenny, p Vince Russo. i Software Systems Research Group University of Illinois at Urbana-Champaign Department of Computer Science 1304 West Springfield Avenue Urbana, Illinois 61801-2987 (217) 333-0215 TABLE OF CONTENTS 1. Introduction. ........................................................................................................................... 1 2. Choices .................................................................................................................................... 1 3. CLASP .................................................................................................................................... 2 4. Path Pascal Release ................................................................................................................. 4 5. The Choices Interface Compiler ................................................................................................ 4 8. Summary ................................................................................................................................. 5 ABSTRACT: Project EOS is studying the problems
    [Show full text]
  • An HDF-EOS and Data Formatting Primer for the ECS Project
    175-WP-001-002 An HDF-EOS and Data Formatting Primer for the ECS Project White Paper March 2001 Prepared Under Contract NAS5-60000 RESPONSIBLE ENGINEER E. M. Taaheri /s/ 3-1-2001 David Wynne, Abe Taaheri Date EOSDIS Core System Project SUBMITTED BY Larry Klein /s/ 2-28-2001 Darryl Arrington, Manager Toolkit Date Larry Klein, Manager Toolkit EOSDIS Core System Project Raytheon Company Upper Marlboro, Maryland This page intentionally left blank. Abstract This document is a release of a primer manual for using the Hierarchical Data Format (HDF) and extensions (HDF-EOS) in Earth Observing System (EOS), Version 1 and later. We confine our discussion in this document to HDF V4 and the versions of HDF-EOS (V2), which are based on that version of HDF. A new version of HDF (V5) has been released. Additional material for HDF5 and versions of HDF-EOS, based on HDF5 will be provided in a separate document. Comments and suggestions should be sent to: Larry Klein Emergent Information Technologies 9315 Largo Dr. West, Suite 250 Largo, MD 20774 USA Email: [email protected] Phone: (301) 925-0764 Fax: (301) 925-0321 Keywords: HDF, HDF-EOS, Swath, Grid, Point, Data Formats, Metadata, Standard Data Products, Disk Formats, Browse iii 175-WP-001-002 This page intentionally left blank. iv 175-WP-001-002 Table of Contents Abstract 1. Introduction 1.1 Purpose of this Document.......................................................................................... 1-1 1.2 Organization of this Document .................................................................................. 1-1 1.3 HDF and EOS: Introduction ...................................................................................... 1-1 1.4 HDF and EOS: General Philosophy........................................................................... 1-3 2.
    [Show full text]
  • Containers: a Sound Basis for a True Single System Image
    Containers : A Sound Basis For a True Single System Image Renaud Lottiaux, Christine Morin To cite this version: Renaud Lottiaux, Christine Morin. Containers : A Sound Basis For a True Single System Image. [Research Report] RR-4085, INRIA. 2000. inria-00072548 HAL Id: inria-00072548 https://hal.inria.fr/inria-00072548 Submitted on 24 May 2006 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. INSTITUT NATIONAL DE RECHERCHE EN INFORMATIQUE ET EN AUTOMATIQUE Containers : A Sound Basis For a True Single System Image Renaud Lottiaux, Christine Morin N˚4085 Novembre 2000 THÈME 1 apport de recherche ISRN INRIA/RR--4085--FR+ENG ISSN 0249-6399 Containers : A Sound Basis For a True Single System Image Renaud Lottiaux , Christine Morin Thème 1 — Réseaux et systèmes Projet PARIS Rapport de recherche n˚4085 — Novembre 2000 — 19 pages Abstract: Clusters of SMPs are attractive for executing shared memory parallel appli- cations but reconciling high performance and ease of programming remains an open issue. A possible approach is to provide an efficient Single System Image (SSI) operating system giving the illusion of an SMP machine. In this paper, we introduce the concept of container as a mechanism to unify global resource management at the lowest operating system level.
    [Show full text]
  • A Single System Image Java Operating System for Sensor Networks
    A SINGLE SYSTEM IMAGE JAVA OPERATING SYSTEM FOR SENSOR NETWORKS Emin Gun Sirer Rimon Barr John C. Bicket Daniel S. Dantas Computer Science Department Cornell University Ithaca, NY 14853 {egs, barr, bicket, ddantas}@cs.cornell.edu Abstract In this paper we describe the design and implementation of a distributed operating system for sensor net- works. The goal of our system is to extend total system lifetime through power-aware adaptation for sensor networking applications. Our system achieves this goal by providing a single system image of a unified Java virtual machine to applications over an ad hoc collection of heterogeneous sensors. It automatically and transparently partitions applications into components and dynamically finds a placement of these components on nodes within the sensor network to reduce energy consumption and increase system longevity. This paper describes the design and implementation of our system and examines the question of where and when to mi- grate components in a sensor network. We evaluate two practical, power-aware, general-purpose algorithms for object placement, as well as an adaptive scheme for deciding the time granularity of object migration. We demonstrate that our algorithms can increase sensor network longevity by a factor of four to five by effec- tively distributing energy consumption and avoiding hotspots. 1. Introduction able to components at each node, in particular the available power and bandwidth may change over Sensor networks simultaneously promise a radi- time and necessitate the relocation of application cally new class of applications and pose signifi- components. Further, event sources that are being cant challenges for application development.
    [Show full text]
  • Architectural Review of Load Balancing Single System Image
    Journal of Computer Science 4 (9): 752-761, 2008 ISSN 1549-3636 © 2008 Science Publications Architectural Review of Load Balancing Single System Image Bestoun S. Ahmed, Khairulmizam Samsudin and Abdul Rahman Ramli Department of Computer and Communication Systems Engineering, University Putra Malaysia, 43400 Serdang, Selangor, Malaysia Abstract: Problem statement: With the growing popularity of clustering application combined with apparent usability, the single system image is in the limelight and actively studied as an alternative solution for computational intensive applications as well as the platform for next evolutionary grid computing era. Approach: Existing researches in this field concentrated on various features of Single System Images like file system and memory management. However, an important design consideration for this environment is load allocation and balancing that is usually handled by an automatic process migration daemon. Literature shows that the design concepts and factors that affect the load balancing feature in an SSI system are not clear. Result: This study will review some of the most popular architecture and algorithms used in load balancing single system image. Various implementations from the past to present will be presented while focusing on the factors that affect the performance of such system. Conclusion: The study showed that although there are some successful open source systems, the wide range of implemented systems investigated that research activity should concentrate on the systems that have already been proposed and proved effectiveness to achieve a high quality load balancing system. Key words: Single system image, NOWs (network of workstations), load balancing algorithm, distributed systems, openMosix, MOSIX INTRODUCTION resources transparently irrespective of where they are available[1].The load balancing single system image Cluster of computers has become an efficient clusters dominate research work in this environment.
    [Show full text]
  • SSI-OSCAR: a Distribution for High Performance Computing Using a Single System Image
    1 SSI-OSCAR: a Distribution For High Performance Computing Using a Single System Image Geoffroy Vallée (INRIA / ORNL / EDF), Christine Morin (INRIA), Stephen L. Scott (ORNL), Jean-Yves Berthou (EDF), Hugues Prisker (EDF) OSCAR Symposium, May 2005 2 Context • Clusters: distributed architecture • difficult to use • difficult to manage • Different approaches • do everything manually • use software suite to simplify management and use (e.g. OSCAR) • This solution does not completely hide the resources distribution • use a Single System Image (SSI) • all resources are managed at the cluster scale • transparent for users and administrators • gives the illusion that a cluster is an SMP machine 3 What is a Single System Image? • SSI features ● Transparent resource management at the cluster level ● High Availability: tolerate all undesirable events that can occurs (node failure or eviction) ● Support of programming standards (e.g. MPI, OpenMP) ● High performance • A Solution: merge OSCAR and an SSI ● Simple to install ● Simple to use ● Collaboration INRIA / EDF / ORNL 4 SSI - Implementation • Key point: global resource management Limitations for functionalities User level: middle-ware and efficiency (e.g. CONDOR) Complex to develop Kernel level: OS and maintain (e.g. MOSIX, OpenSSI, Kerrighed) Hardware level More expensive (e.g. SGI) 5 Kerrighed – Overview • SSI developed in France (Rennes), INRIA/IRISA, in collaboration with EDF • Management at the cluster scale of • memories (through a DSM) • processes (through mechanisms for global process
    [Show full text]
  • The Intel Paragon
    Operating system and network support for high-performance computing Item Type text; Dissertation-Reproduction (electronic) Authors Guedes Neto, Dorgival Olavo Publisher The University of Arizona. Rights Copyright © is held by the author. Digital access to this material is made possible by the University Libraries, University of Arizona. Further transmission, reproduction or presentation (such as public display or performance) of protected items is prohibited except with permission of the author. Download date 30/09/2021 20:24:35 Link to Item http://hdl.handle.net/10150/298757 INFORMATION TO USERS This manuscript has been reproduced from the microfilm master. UMI films the text directly from the original or copy submitted. Thus, some thesis and dissertation copies are in typewriter face, while others may be from any type of computer printer. The quality of this reproduction is dependent upon the quality of the copy submitted. Broken or indistinct print, colored or poor quality illustrations and photographs, print bleedthrough, substandard margins, and improper alignment can adversely affect reproduction. In the unlikely event that the author did not send UMI a complete manuscript and there are missing pages, these will be noted. Also, if unauthorized copyright material had to be removed, a note will indicate the deletion. Oversize materials (e.g., maps, drawings, charts) are reproduced by sectioning the original, beginning at the upper left-hand comer and continuing from left to right in equal sections with small overlaps. Each original is also photographed in one exposure and is included in reduced form at the back of the book. Photographs included in the original manuscript have been reproduced xerographically in this copy.
    [Show full text]
  • Quantian: a Single-System Image Scientific Cluster Computing
    Quantian: A single-system image scientific cluster computing environment Dirk Eddelbuettel, Ph.D. B of A, and Debian [email protected] Presentation at the Extreme Linux SIG at USENIX 2004 in Boston, July 2, 2004 Quantian: A single-system image scientific cluster computing environment – p. 1 Introduction Quantian is a directly bootable and self-configuring Linux sytem that runs from a compressed dvd image. Quantian offers zero-configuration cluster computing using openMosix. Quantian can boot ’thin clients’ directly via PXE in an ’openmosixterminalserver’ setting. Quantian contains around 1gb of additional ’quantitative’ software: scientific, numerical, statistical, engineering, ... Quantian also contains tools of general usefulness such as editors, programming languages, a very complete latex suite, two ’office’ suites, networking tools and multimedia apps. Quantian: A single-system image scientific cluster computing environment – p. 2 Family tree overview Quantian is based on clusterKnoppix, which extends Knoppix with an openMosix-enabled kernel and applications (chpox, gomd, tyd, ....), kernel modules and security patches. ClusterKnoppix extends Knoppix, an impressive ’linux on a cdrom’ system which puts 2.3gb of software onto a cdrom along with the very best auto-detection and configuration. Knoppix is based on Debian, a Linux distribution containing over 6000 source packages available for 10 architectures (such as i386, alpha, ia64, amd64, sparc or s390) produced by hundreds of individuals from across the globe. Quantian: A single-system image scientific cluster computing environment – p. 3 Family tree: Debian ’Linux the Linux way’: made by volunteers (some now with full-time backing) from across the globe. Focus on very high technical standards with rigorous policy and reference documents.
    [Show full text]