Digital Technical Journal, Number 5, September 1987: Vaxcluster Systems

Total Page:16

File Type:pdf, Size:1020Kb

Digital Technical Journal, Number 5, September 1987: Vaxcluster Systems VAX clusterSystems Digital Technical Journal Digital Equipment Corporation Number 5 September 1987 Editorial Staff Ediwr- Richard W. Beane Production Staff Production Editor- Jane C. 131ake Designer- Charlotte Bell Interactive Page Makeup- Te rry Reed Advisory Board Samuel H. Fuller, Chairman Robert M. Glorioso John W. McCredie Mahendra R. Patel F. Grant Saviers William D. Strecker The Digital Te chnical journal is published by Digital Equipment Corporation, 77 Reed Road, Hudson, MassachusettS 01749. Changes of address should be sent to Digital Equipment Corporation, attention: Media Response Manager, 444 Whitney Street, NR02-1/J5, Northboro, M.A 01532-2599 Comments on the content of any paper are welcomed. Write to the editor at Mail Stop HL02-3/K11 at the published-by address. Comments can also be sent on the ENET to RDVAX::BEANE or on the ARPANET to BEANE%RDVAX.DEC@DECWRL. Copyright© 1987 Digital Equipment Corporation. Copying without fee is permiued provided that such copies are made for use in educational institutions by facuhy members and are not distributed for commercial advantage. Abstracting with credit of Digital Equipment Corporation's authorship is permitted Requests for other copies for a fee may be made to the Digital Press of Digital Equipment Corporation. All rights reserved. The information in this journal is subject to change without notice and should not be construed as a commitment by Digital Equipment Corporation. Digital Equipment Corporation assumes no responsibility for any errors that may appear in this document. ISBN 1-55558-004-1 Documentation Number EY-8258E-DP The following are trademarks of Digital Equipment Corporation: CI, DEC, DECnet, DECnet-VAX, DECsystem-1 0, DECSYSTEM-20, Digital Network Architecture (DNA), Digital Storage Architecture (DSA), the Digital logo, HSC, Local Area VAXcluster, MicroVAX, MicroVAX II, MicroVAX 2000, Q-bus, RJ\1S- II, SA482, UNIBUS, VAX, VAX-11/750, VAX-11/780, VAX-11/782, VAX-11/785, VAX 8600, VAX 8650, VAX 8700, VAX 8974, VAX 8978, VAXcluster, VAXstation, VAXstation JJ, VAXstation ll/GPX, VA.Xstation 2000, VMS, Cover Design VT, VT220 VAX cluster systems are featured in this issue. The central connection between the elements in a cluster is called the IBM is a registered trademark of International Business Machines, Inc. Star Coupler. Our star-filled cover evokes the thousands of VAXcluster systems now operating worldwide. The image was Intel is a trademark of Intel Corporation. created using the Lightspeed System. Lightspeed is a trademark of Lightspeed Computers, Inc. The cover was designed by Barbara Grzeslo and Tim Roberts Book production was done by Educational Services of the Graphic Design Department. Media Communications Group in Bedford, MA. Contents V AXcluster Systems 7 The VAX cluster Concept: An Overview of a Distributed System Nancy P. Kronenberg, Henry M. Levy, William D. Strecker, and Richard J. Merewood 22 The System Communication Architecture Darrell J. Duffy 29 The VAX/VMS Distributed Lock Manager William E. Snaman, Jr. and David W. Thiel 4 5 The Design and Implementation of a Distributed File System Andrew C. Goldstein 56 Local Area VAXcluster Systems Michael S. Fox and John A. Ywoskus 69 VAXcluster Availability Modeling Edward E. Balkovich, Prashant Bhabhalia, William R. Dunnington, and Thomas F. Weyant 80 System Level Performance of VAX 8974 and 8978 Systems Daeil Park, Rekha D. Von Ehren, Tzyh-jong Wang, and Nii N. Quaynor 93 C/ Bus Arbitration Performance in a VAXcluster System Xi-ren Cao, Nii N. Quaynor, and Fernando C. Colon Osorio Editor:ls Introduction Additional fe atures were needed in the VMS software to accommodate accessing disks on multiple systems. The distributed lock manager, described by Sandy Snaman and Dave Thiel, pro· vides the synchronization needed to accomplish transparent data transfers between cluster mem· hers. Other changes were also needed to broaden the file functions performed by the VMS software. Andy Goldstein relates some alternative ways to expand those functions and how the QJO proces· sor was extended ro synchronize file accesses. The resulting system of locks and queues pro· Richal"d W. Beane Editor vides a consistent sequence for managing dis­ tributed files. The next paper. by Mike Fox and john Ywoskus, describes the extension of the VA.-'Ccluster con­ VAXcluster systems are closely coupled configu· cept ro systems connected with an Ethernet. rations of VAX CPUs and storage devices. The VAX Tbcsc Local Area VAXcluster systems use special CPU at any node can comm unicate with software to provide fu nctions needed by clusters, the processor and storage devices at any other but not provided by Ethernet software. Thus, node in the cluster. The interconnects and soft· MicroVAX II and other small VAX systems can be ware used to activate this unique concept allow clustered to yield significant amounts of process· data transfers at up to 70 megabits per second ing power. between nodes. This issue of the Digital Te chni· The last three papers deal with performance cal jo urnal contains papers about some of the aspects of VAXcluster systems. The paper by Ed key hardware and software features in these sys· Ba lkovich, Prashant Bhabhalia, Dick Dunnington, terns , as well as some measures of their perfor· and Tom Weyant discusses the results of a VA X· mance. Since several organizations within Digital cluster model that demonstrates how redundancy are responsible for various VA Xcluster features , improves availability. Then, Dale Pa rk, Re kha these papers are comribured by engineers from a Von Ehren, T·J. Wa ng, and Nii Quaynor describe wide spectrum of engineering groups. two models they developed to measure the per· Since the VAXc luster concept spans such a formances of VAX 8974 and 8978 systems. These range of technologies, the first paper is an models, based on benchmarks run in different overview explaining generally how these sys­ environments, use a VAX 8700 CPU for a baseline tems work. Nancy Kronenberg, Hank Levy, Bill comparison. Strecker, and Richard Merewood describe the The fi nal paper relates the resu lts of a model to architecture, the storage control , the VMS soft· measure the characteristics of the CI bus. Xi-rcn ware alterations, and the multitude of activities Cao , Nii Quaynor, and Fernando Colon Osorio that control access to the storage devices. describe how their model measures the per· The System Communication Architecture, fo rmance of the arbitration algorithm in this described by Darrell Duffy , is the structure that bus. They suggest some interesting schemes to allows the nodes in a VA Xcluster system to coop­ improve utilization and reduce response time. erate. This relatively simple framework governs the sharing of data between resources at the nodes and binds together applications that run on diffe rent VAX CPUs. 2 Biographies Edward E. Balkovich Ed Ba lkov ich is the manager of V�'{c luster System Engineering, which addresses issues of VAXcluster pcrformance, availability and architecture for High Performance Systems. He was Digital's associate director of Project Athena at M.I.T. and is an Adjunct A.o;sociate Professor at Brandeis University. Before joining Digital in 1 9RI, Eel was a fa culty member at the University of Connecticut. He earned his B.A. degree ( 1968) from the University of California at Berkeley, and his M.S. ( 197 1) and Ph. D. (1 976) degrees from the University of California at Santa Ba rbara . He is a member of the ACM and IEEE. Prashant Bhabhalia A principal engineer in V�'Ccluster Systems Engi· neering, Prashant Bhabhalia develops and interprets re liability and availabil· ity models. Earlier, he was a program manager in Computer Systems Manu­ facturing and a senior engineer in GlA Manufacturing. Before joining Digital in 19RO, Prashant was an industrial engineer at Norton Company and Gits Plastic Corporation. He holds an M.S. I.E. degree ( 1974) from the Polytech­ nic Institute of Brooklyn and a B.S.M.E. degree ( 1972) from the M.S. Univer­ sity in India. Prashant is a senior member of l.I.E. Xi-Ren Cao A.<> a principal software engineer in the High Performance Sys­ tems and Clusters Group, Xi-Ren Cao models and evaluates VAX cluster con­ figurations. Before joining Digital in 1986, he was a research fe llow at Har­ vard Universiry. Xi-Ren has published over 20 technical papers on performance evaluation, si mulation, srochastic systems, queuing networks, and control theory, and has co-authored a book "Perrurbation Analysis of Dis­ crete Event Systems," to be published in 1988. He received his Ph.D. degree from Harvard University in 19R4 and is a member of IEEE. Fernando C. Colon Osorio Fernando Colon Osorio graduated from the University of Puerto Rico (B.S.E.E., 1970) and the University of Massachu­ setts (M.S., Ph.D , 1976). Joining Digital in l97(J. he helped design the PDP- 1Ij60 and PDP-11/74 systems and managed the LAN group in Corporate Research. Fernando also managed the overall design verification for the V�'{ 8600 project. In High Performance Systems. he now manages the systems research and advanced development group, responsible for VAXclusters. fault tolerance, advanced architectures, and performance analyses. He was Associate Editor of the IEEE Transactions on Computers and is the co-author of "Engineering Intelligent Systems." Darrellj. Duffy As a consulting software engineer. Dan-ell Duffy works on the network architecture for VAXcluster systems. On previous projects, he led the development of operating systems for parallel processors and wrote software for the Local Area Te rminal protocol . Darrell helped to deve lop DECnet software after joining Digital in 1977 He received a B.S. in com­ puter science from West Virginia University in 1972 and worked at the Un i­ versity of Florida. Darrell and three other Digital engineers have applied fo r a parent on the LAT protocol .
Recommended publications
  • Openvms Record Management Services Reference Manual
    OpenVMS Record Management Services Reference Manual Order Number: AA-PV6RD-TK April 2001 This reference manual contains general information intended for use in any OpenVMS programming language, as well as specific information on writing programs that use OpenVMS Record Management Services (OpenVMS RMS). Revision/Update Information: This manual supersedes the OpenVMS Record Management Services Reference Manual, OpenVMS Alpha Version 7.2 and OpenVMS VAX Version 7.2 Software Version: OpenVMS Alpha Version 7.3 OpenVMS VAX Version 7.3 Compaq Computer Corporation Houston, Texas © 2001 Compaq Computer Corporation Compaq, AlphaServer, VAX, VMS, the Compaq logo Registered in U.S. Patent and Trademark Office. Alpha, PATHWORKS, DECnet, DEC, and OpenVMS are trademarks of Compaq Information Technologies Group, L.P. in the United States and other countries. UNIX and X/Open are trademarks of The Open Group in the United States and other countries. All other product names mentioned herein may be the trademarks of their respective companies. Confidential computer software. Valid license from Compaq required for possession, use, or copying. Consistent with FAR 12.211 and 12.212, Commercial Computer Software, Computer Software Documentation, and Technical Data for Commercial Items are licensed to the U.S. Government under vendor’s standard commercial license. Compaq shall not be liable for technical or editorial errors or omissions contained herein. The information in this document is provided "as is" without warranty of any kind and is subject to change without notice. The warranties for Compaq products are set forth in the express limited warranty statements accompanying such products. Nothing herein should be construed as constituting an additional warranty.
    [Show full text]
  • Software Product Description and Quickspecs
    VSI OpenVMS Alpha Version 8.4-2L2 Operating System DO-DVASPQ-01A Software Product Description and QuickSpecs PRODUCT NAME: VSI OpenVMS Alpha Version 8.4-2L2 DO-DVASPQ-01A This SPD and QuickSpecs describes the VSI OpenVMS Alpha Performance Release Operating System software, Version 8.4-2L2 (hereafter referred to as VSI OpenVMS Alpha V8.4-2L2). DESCRIPTION OpenVMS is a general purpose, multiuser operating system that runs in both production and development environments. VSI OpenVMS Alpha Version 8.4-2L2 is the latest release of the OpenVMS Alpha computing environment by VMS Software, Inc (VSI). VSI OpenVMS Alpha V8.4-2L2 is compiled to take advantage of architectural features such as byte and word memory reference instructions, and floating-point improvements, which are available only in HPE AlphaServer EV6 or later processors. This optimized release improves performance by taking advantage of faster hardware-based instructions that were previously emulated in software. NOTE: VSI OpenVMS Alpha V8.4-2L2 does not work on, and is not supported on, HPE AlphaServer pre-EV6 systems. OpenVMS Alpha supports HPE’s AlphaServer series computers. OpenVMS software supports industry standards, facilitating application portability and interoperability. OpenVMS provides symmetric multiprocessing (SMP) support for multiprocessing systems. The OpenVMS operating system can be tuned to perform well in a wide variety of environments. This includes combinations of compute-intensive, I/O-intensive, client/server, real-time, and other environments. Actual system performance depends on the type of computer, available physical memory, and the number and type of active disk and tape drives. The OpenVMS operating system has well-integrated networking, distributed computing, client/server, windowing, multi-processing, and authentication capabilities.
    [Show full text]
  • Operating System Support for Parallel Processes
    Operating System Support for Parallel Processes Barret Rhoden Electrical Engineering and Computer Sciences University of California at Berkeley Technical Report No. UCB/EECS-2014-223 http://www.eecs.berkeley.edu/Pubs/TechRpts/2014/EECS-2014-223.html December 18, 2014 Copyright © 2014, by the author(s). All rights reserved. Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, to republish, to post on servers or to redistribute to lists, requires prior specific permission. Operating System Support for Parallel Processes by Barret Joseph Rhoden A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in Computer Science in the Graduate Division of the University of California, Berkeley Committee in charge: Professor Eric Brewer, Chair Professor Krste Asanovi´c Professor David Culler Professor John Chuang Fall 2014 Operating System Support for Parallel Processes Copyright 2014 by Barret Joseph Rhoden 1 Abstract Operating System Support for Parallel Processes by Barret Joseph Rhoden Doctor of Philosophy in Computer Science University of California, Berkeley Professor Eric Brewer, Chair High-performance, parallel programs want uninterrupted access to physical resources. This characterization is true not only for traditional scientific computing, but also for high- priority data center applications that run on parallel processors. These applications require high, predictable performance and low latency, and they are important enough to warrant engineering effort at all levels of the software stack.
    [Show full text]
  • Understanding Lustre Filesystem Internals
    ORNL/TM-2009/117 Understanding Lustre Filesystem Internals April 2009 Prepared by Feiyi Wang Sarp Oral Galen Shipman National Center for Computational Sciences Oleg Drokin Tom Wang Isaac Huang Sun Microsystems Inc. DOCUMENT AVAILABILITY Reports produced after January 1, 1996, are generally available free via the U.S. Department of Energy (DOE) Information Bridge. Web site http://www.osti.gov/bridge Reports produced before January 1, 1996, may be purchased by members of the public from the following source. National Technical Information Service 5285 Port Royal Road Springfield, VA 22161 Telephone 703-605-6000 (1-800-553-6847) TDD 703-487-4639 Fax 703-605-6900 E-mail [email protected] Web site http://www.ntis.gov/support/ordernowabout.htm Reports are available to DOE employees, DOE contractors, Energy Technology Data Exchange (ETDE) representatives, and International Nuclear Information System (INIS) representatives from the following source. Office of Scientific and Technical Information P.O. Box 62 Oak Ridge, TN 37831 Telephone 865-576-8401 Fax 865-576-5728 E-mail [email protected] Web site http://www.osti.gov/contact.html This report was prepared as an account of work sponsored by an agency of the United States Government. Neither the United States Government nor any agency thereof, nor any of their employees, makes any warranty, express or implied, or assumes any legal liability or responsibility for the accuracy, completeness, or usefulness of any information, apparatus, product, or process disclosed, or represents that its use would not infringe privately owned rights. Reference herein to any specific commercial product, process, or service by trade name, trademark, manufacturer, or otherwise, does not necessarily constitute or imply its endorsement, recommendation, or favoring by the United States Government or any agency thereof.
    [Show full text]
  • High Availability for RHEL on System Z
    High Availability for RHEL on System z David Boyes Sine Nomine Associates Agenda • Clustering • High Availability • Cluster Management • Failover • Fencing • Lock Management • GFS2 Clustering • Four types – Storage – High Availability – Load Balancing – High Performance High Availability • Eliminate Single Points of Failure • Failover • Simultaneous Read/Write • Node failures invisible outside the cluster • rgmanager is the core software High Availability • Major Components – Cluster infrastructure — Provides fundamental functions for nodes to work together as a cluster • Configuration-file management, membership management, lock management, and fencing – High availability Service Management — Provides failover of services from one cluster node to another in case a node becomes inoperative – Cluster administration tools — Configuration and management tools for setting up, configuring, and managing the High Availability Implementation High Availability • Other Components – Red Hat GFS2 (Global File System 2) — Provides a cluster file system for use with the High Availability Add-On. GFS2 allows multiple nodes to share storage at a block level as if the storage were connected locally to each cluster node – Cluster Logical Volume Manager (CLVM) — Provides volume management of cluster storage – Load Balancer — Routing software that provides IP-Load-balancing Cluster Infrastructure • Cluster management • Lock management • Fencing • Cluster configuration management Cluster Management • CMAN – Manages quorum and cluster membership – Distributed
    [Show full text]
  • Comparative Analysis of Distributed and Parallel File Systems' Internal Techniques
    Comparative Analysis of Distributed and Parallel File Systems’ Internal Techniques Viacheslav Dubeyko Content 1 TERMINOLOGY AND ABBREVIATIONS ................................................................................ 4 2 INTRODUCTION......................................................................................................................... 5 3 COMPARATIVE ANALYSIS METHODOLOGY ....................................................................... 5 4 FILE SYSTEM FEATURES CLASSIFICATION ........................................................................ 5 4.1 Distributed File Systems ............................................................................................................................ 6 4.1.1 HDFS ..................................................................................................................................................... 6 4.1.2 GFS (Google File System) ....................................................................................................................... 7 4.1.3 InterMezzo ............................................................................................................................................ 9 4.1.4 CodA .................................................................................................................................................... 10 4.1.5 Ceph.................................................................................................................................................... 12 4.1.6 DDFS ..................................................................................................................................................
    [Show full text]
  • The Chubby Lock Service for Loosely-Coupled Distributed Systems
    The Chubby lock service for loosely-coupled distributed systems Mike Burrows, Google Inc. Abstract example, the Google File System [7] uses a Chubby lock to appoint a GFS master server, and Bigtable [3] uses We describe our experiences with the Chubby lock ser- Chubby in several ways: to elect a master, to allow the vice, which is intended to provide coarse-grained lock- master to discover the servers it controls, and to permit ing as well as reliable (though low-volume) storage for clients to find the master. In addition, both GFS and a loosely-coupled distributed system. Chubby provides Bigtable use Chubby as a well-known and available loca- an interface much like a distributed file system with ad- tion to store a small amount of meta-data; in effect they visory locks, but the design emphasis is on availability use Chubby as the root of their distributed data struc- and reliability, as opposed to high performance. Many tures. Some services use locks to partition work (at a instances of the service have been used for over a year, coarse grain) between several servers. with several of them each handling a few tens of thou- Before Chubby was deployed, most distributed sys- sands of clients concurrently. The paper describes the tems at Google used ad hoc methods for primary elec- initial design and expected use, compares it with actual tion (when work could be duplicated without harm), or use, and explains how the design had to be modified to required operator intervention (when correctness was es- accommodate the differences.
    [Show full text]
  • Systemprogrammierung Grundlage Von Betriebssystemen
    Systemprogrammierung Grundlage von Betriebssystemen Sachwortverzeichnis c Wolfgang Schr¨oder-Preikschat 27. Januar 2017 Die in der Vorlesung (mundlich¨ oder schriftlich) verwendeten Akronyme und Sachworte sind in der nachfolgenden Aufz¨ahlung zusammengefasst. Dabei werden die Akronyme nach den Sachwor- ten, fur¨ die sie die Verkurzung¨ bilden, aufgeschlusselt.¨ Fur¨ englischsprachige Sachw¨orter werden, soweit bekannt, die deutschsprachigen Entsprechungen angegeben. In der Beschreibung durch das "-Zeichen angefuhrte¨ Sachw¨orter zeigen einen Kreuzverweis an. Jedes Sachwort wird erkl¨art, wo- bei dies im Zusammenhang mit dem hier relevanten Kontext der "Systemprogrammierung und in Bezug auf Betriebssysteme geschieht. Die Formulierungen erheben nicht den Anspruch auf Gultigkeit¨ auch fur¨ andere Fachrichtungen in der Informatik. Ebenso erhebt die Aufz¨ahlung nicht den Anspruch auf Vollst¨andigkeit fur¨ das in der Vorlesung behandelte Fachgebiet. Der vorliegende Text ist ein Nachschlagewerk, dessen Lekture,¨ im Gegensatz zu einem Lehr- buch mit durchgehendem roten Faden, fur¨ gew¨ohnlich nicht von vorne nach hinten empfohlen ist. Vielmehr ist der Text Begleitmaterial zu den Vorlesungsfolien. Neben Systemprogrammierung sind hier insbesondere die Lehrveranstaltungen Betriebssysteme, Betriebssystemtechnik, Nebenl¨aufige Systeme und Echtzeitsysteme eingeschlossen. Daruber¨ hinaus kommen aber auch Lehrveranstal- tungen zu den Themen Rechnerorganisation und Rechnerarchitektur als Bezugspunkt in Frage. Viele der dort (mundlich¨ oder schriftlich)
    [Show full text]
  • Proceedings, ITC/USA
    THE EFFICIENT USE OF A VAX COMPUTER IN THE REAL-TIME TELEMETRY ENVIRONMENT Item Type text; Proceedings Authors Robbins, Robert B. Publisher International Foundation for Telemetering Journal International Telemetering Conference Proceedings Rights Copyright © International Foundation for Telemetering Download date 01/10/2021 10:32:25 Link to Item http://hdl.handle.net/10150/613377 THE EFFICIENT USE OF A VAX COMPUTER IN THE REAL-TIME TELEMETRY ENVIRONMENT Robert B. Robbins Sangamo-Weston Incorporated Data Systems Division Post Office Box 3041 Sarasota, Florida 33578 ABSTRACT The use of a Digital Equipment Corporation VAX computer under the VMS operating system, in a real-time telemetry environment, brings with it many advantages. These advantages pertain to its ability to handle real-time telemetry processing in an efficient and relatively straight forward manner. The author will use the TELSET, TELDAX and TELFOR telemetry software systems as the basis for demonstrating the techniques which have allowed the real-time telemetry user to take advantage of a 32-bit, virtual addressing, architecture. INTRODUCTION Many real-time telemetry computer users believe in the need for 32-bit processing of telemetry data once it has entered the central processor. We at Sangamo-Weston agree with this belief and have successfully designed a real-time telemetry data acquisition system which efficiently uses the VAX computer under the VMS operating system. SYSTEM SOFTWARE The software written for this system achieves three goals. TELSET, which sets up the telemetry front-end hardware units, is written in VAX-11 MACRO. TELDAX the real- time data acquisition executive and TELFOR the telemetry data acquisition formatted storage process are written in interactive VAX-11 FORTRAN and are menu driven.
    [Show full text]
  • RSX - 11 M-PLUS Mini-Reference
    RSX - 11 M-PLUS Mini-Reference Order No. AV-H435F-TC RSX - 11M-PLUS Mini-Reference Order Number. AV-H435F-TC RSX-ll M-PLUS Version 4.2 Digital Equipment Corporation Maynard, Massachusetts First Printing, September 1977 Revised, April 1982 Revised, April 1983 Revised, July 1985 R~vised, .September 1987 Revised, January 1989 The information in this document is subject to change without notice and should not be construed as a commitment by Digital Equipment Corporation. Digital Equipment Corporation assumes no responsibility for any errors that may appear in this document. The software described in this document is furnished under a license and may be used or copied only in accordance with the terms of such license. No responsibility is assumed for the use or reliability of software on equipment that is not supplied by Digital Equipment Corporation or its affiliated companies. © Digital Equipment Corporation 1977, 1982, 1983, 1985, 1987, 1989. All Rights Reserved. Printed in U.S.A. The postpaid Reader's Comments forms at the end of this document request your critical evaluation to assist in preparing future documentation. The following are trademarks of Digital Equipment Corporation: DEC DIBOL UNIBUS DEC/CMS EduSystem VAX DEC/MMS lAS VAXcluster DECnet MASSBUS VMS DECsystem-lO PDP VT DECSYSTEM-20 PDT DECUS RSTS DECwriter RSX ~U~UIl~DTM ZK5077 Contents Preface vii Conventions ............................................... viii Online Help Files Online Help Files ............................................. 3 Command Line Interpreters Monitor Console Routine (MCR) Commands ......................... 7 Digital Command Language (DCL) ............................... 21 utilities BAD Command Summary ...................................... 67 iii BRU Command Summary ...................................... 69 CMP Command Summary ...................................... 74 DMP Command Summary .....................................
    [Show full text]
  • Scaling HDFS with a Strongly Consistent Relational Model for Metadata Kamal Hakimzadeh, Hooman Peiro Sajjad, Jim Dowling
    Scaling HDFS with a Strongly Consistent Relational Model for Metadata Kamal Hakimzadeh, Hooman Peiro Sajjad, Jim Dowling To cite this version: Kamal Hakimzadeh, Hooman Peiro Sajjad, Jim Dowling. Scaling HDFS with a Strongly Consistent Relational Model for Metadata. 4th International Conference on Distributed Applications and In- teroperable Systems (DAIS), Jun 2014, Berlin, Germany. pp.38-51, 10.1007/978-3-662-43352-2_4. hal-01287731 HAL Id: hal-01287731 https://hal.inria.fr/hal-01287731 Submitted on 14 Mar 2016 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Distributed under a Creative Commons Attribution| 4.0 International License Scaling HDFS with a Strongly Consistent Relational Model for Metadata Kamal Hakimzadeh, Hooman Peiro Sajjad, Jim Dowling KTH - Royal Institute of Technology Swedish Institute of Computer Science (SICS) {mahh, shps, jdowling}@kth.se Abstract. The Hadoop Distributed File System (HDFS) scales to store tens of petabytes of data despite the fact that the entire le system's metadata must t on the heap of a single Java virtual machine. The size of HDFS' metadata is limited to under 100 GB in production, as garbage collection events in bigger clusters result in heartbeats timing out to the metadata server (NameNode).
    [Show full text]
  • Clustering of Openvms Installations for High Availability
    Clustering of OpenVMS installations for high availability Norman Kluge norman.kluge [at] student.hpi.uni-potsdam.de Hasso-Plattner-Institut, Potsdam Lecture: Dependable Systems Summer term 2010 Key words: OpenVMS, Cluster, Availabilty, Dependable Systems 1 Introduction Nowadays services have to be high available. Datacenters are often fault tolerant. That means they are for instance tolerant against hardware failures, software fail- ures or electricity failures. But what happened if a hole datacenter is destroyed. High availabilty means that the service still has to be online. OpenVMSCluster oers concepts of so called disaster tolerance, means even if a datacenter fails, the service never stops the beat. That is ensured by dierent approaches covered in this paper. In the rst section, some terminology of OpenVMSCluser are introduced. In the second one, some concepts, how OpenVMSCluser ensures high availabilty, are covered. In the las section, some practical experience is mentioned. 2 OpenVMSCluster Basics In this section some basic terminology of OpenVMSCluster is described. 2.1 Benets There are some benets resulting from clustering of OpenVMS installations. First point is, that the dierent Workstations can share resources (e.g. disks, network connections, ...) among them. The sharing concept is very important for another important benet: the promise of high availability. With this and nonstop processing an OpenVMSCluster guarantees that the services running on it are always available and responding the user. That correlates with the benets of scalability, performance and load balancing, which means that the load is spread into the cluster and all available workstations are working on the task. There are also some security features oered by OpenVMSCluster.
    [Show full text]