Frangipani: a Scalable Distributed File System
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
The New Standard for Data Archiving
The new standard for data archiving 1 Explosively increasing 2 Why data management digital data is important Ever-increasing volumes of digital data are mounting up every day due to rapidly growing Today, files can be lost from computers in any number of ways—you might accidentally internet technology, widespread use of SNS, data transmission between network- connected delete a file, a virus might wipe one out, or there could be a complete hard drive failure. devices, and other trends. Within the video production industry, data-heavy video content When a hard drive dies an untimely death, it can feel like a house has burnt down. (for example, 4K, 8K, and 4K/8K high-frame-rate video) is becoming a major source of video Important personal items are usually gone forever—photos, significant documents, broadcasting. Many companies and research institutes are creating high volumes of data downloaded music, and more. (big data) for use in AI systems. Somehow, these newly created assets need to be managed effectively, stored safely, and utilized along with the old assets. There are many options for backing up content without any sophisticated equipment—you can 163,000 EB use DVDs, external hard drives, optical discs, or even online storage. It’s a good idea to back up data to multiple places. Computer Natural Viruses Disasters 4% Other 2% 40,026 EB 3% 2,837 EB 8,591 EB 1,227 EB Software Corruption 9% 2010 2012 2015 2020 2025 System/ Hardware 1 EB = 1,000,000 TB Malfunction 56% User Error Performance 26% Source: Ontrack Data Recovery High Performance www.ontrack.co.uk/understandingdataloss E-commerce, Financial HOT SSD RAM Capacity Optimized Back Oce, General Service WARM HDD Sony's Optical Disc Archive storage system offers the solution, Long-Term Archive COLD with a low total cost of ownership through the use of long-life Auto Loader, O-line Tape Optical Disc Cost media, and it includes inter-generational compatibility based on the same optical disc technology used in DVDs and Blu-ray discs. -
Andrew File System (AFS) Google File System February 5, 2004
Advanced Topics in Computer Systems, CS262B Prof Eric A. Brewer Andrew File System (AFS) Google File System February 5, 2004 I. AFS Goal: large-scale campus wide file system (5000 nodes) o must be scalable, limit work of core servers o good performance o meet FS consistency requirements (?) o managable system admin (despite scale) 400 users in the “prototype” -- a great reality check (makes the conclusions meaningful) o most applications work w/o relinking or recompiling Clients: o user-level process, Venus, that handles local caching, + FS interposition to catch all requests o interaction with servers only on file open/close (implies whole-file caching) o always check cache copy on open() (in prototype) Vice (servers): o Server core is trusted; called “Vice” o servers have one process per active client o shared data among processes only via file system (!) o lock process serializes and manages all lock/unlock requests o read-only replication of namespace (centralized updates with slow propagation) o prototype supported about 20 active clients per server, goal was >50 Revised client cache: o keep data cache on disk, metadata cache in memory o still whole file caching, changes written back only on close o directory updates are write through, but cached locally for reads o instead of check on open(), assume valid unless you get an invalidation callback (server must invalidate all copies before committing an update) o allows name translation to be local (since you can now avoid round-trip for each step of the path) Revised servers: 1 o move -
Restoring in All Situations
Restoring in all situations Desktop and laptop protection 1 Adapt data restore methods to different situations Executive Summary Whether caused by human error, a cyber attack, or a physical disaster, data loss costs organizations millions of euros every year (3.5 million euros on average according to the Ponemon Institute). Backup Telecommuting Office remains the method of choice to protect access to company data and ensure their availability. But a good backup strategy must necessarily be accompanied by a good disaster recovery strategy. Nomadization The purpose of this white paper is to present the best restore practices depending on the situation you are in, in order to save valuable time! VMs / Servers Apps & DBs NAS Laptops Replication 2 Table of Contents EXECUTIVE SUMMARY .......................................................................................................................... 2 CONTEXT ................................................................................................................................................ 4 WHAT ARE THE STAKES BEHIND THE RESTORING ENDPOINT USER DATA? ................................. 5 SOLUTION: RESTORING ENDPOINT DATA USING LINA ................................................................... 6-7 1. RESTORE FOR THE MOST NOVICE USERS ................................................................................ 8-10 2. RESTORE FOR OFF SITE USERS ................................................................................................11-13 3. RESTORE FOLLOWING LOSS, -
IBM Connect:Direct for Microsoft Windows: Documentation Fixpack 1 (V6.1.0.1)
IBM Connect:Direct for Microsoft Windows 6.1 Documentation IBM This edition applies to Version 5 Release 3 of IBM® Connect:Direct and to all subsequent releases and modifications until otherwise indicated in new editions. © Copyright International Business Machines Corporation 1993, 2018. US Government Users Restricted Rights – Use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM Corp. Contents Chapter 1. Release Notes.......................................................................................1 Requirements...............................................................................................................................................1 Features and Enhancements....................................................................................................................... 2 Special Considerations................................................................................................................................ 3 Known Restrictions...................................................................................................................................... 4 Restrictions for Connect:Direct for Microsoft Windows........................................................................ 4 Restrictions for Related Software.......................................................................................................... 6 Installation Notes.........................................................................................................................................6 -
A Survey of Distributed File Systems
A Survey of Distributed File Systems M. Satyanarayanan Department of Computer Science Carnegie Mellon University February 1989 Abstract Abstract This paper is a survey of the current state of the art in the design and implementation of distributed file systems. It consists of four major parts: an overview of background material, case studies of a number of contemporary file systems, identification of key design techniques, and an examination of current research issues. The systems surveyed are Sun NFS, Apollo Domain, Andrew, IBM AIX DS, AT&T RFS, and Sprite. The coverage of background material includes a taxonomy of file system issues, a brief history of distributed file systems, and a summary of empirical research on file properties. A comprehensive bibliography forms an important of the paper. Copyright (C) 1988,1989 M. Satyanarayanan The author was supported in the writing of this paper by the National Science Foundation (Contract No. CCR-8657907), Defense Advanced Research Projects Agency (Order No. 4976, Contract F33615-84-K-1520) and the IBM Corporation (Faculty Development Award). The views and conclusions in this document are those of the author and do not represent the official policies of the funding agencies or Carnegie Mellon University. 1 1. Introduction The sharing of data in distributed systems is already common and will become pervasive as these systems grow in scale and importance. Each user in a distributed system is potentially a creator as well as a consumer of data. A user may wish to make his actions contingent upon information from a remote site, or may wish to update remote information. -
File Server Scaling with Network-Attached Secure Disks
Appears in Proceedings of the ACM International Conference on Measurement and Modeling of Computer Systems (Sigmetrics ‘97), Seattle, Washington, June 15-18, 1997. File Server Scaling with Network-Attached Secure Disks Garth A. Gibson†, David F. Nagle*, Khalil Amiri*, Fay W. Chang†, Eugene M. Feinberg*, Howard Gobioff†, Chen Lee†, Berend Ozceri*, Erik Riedel*, David Rochberg†, Jim Zelenka† *Department of Electrical and Computer Engineering †School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213-3890 [email protected] http://www.cs.cmu.edu/Web/Groups/NASD/ Abstract clients with efficient, scalable, high-bandwidth access to stored data. This paper discusses a powerful approach to fulfilling this By providing direct data transfer between storage and client, net- need. Network-attached storage provides high bandwidth by work-attached storage devices have the potential to improve scal- directly attaching storage to the network, avoiding file server ability for existing distributed file systems (by removing the server store-and-forward operations and allowing data transfers to be as a bottleneck) and bandwidth for new parallel and distributed file striped over storage and switched-network links. systems (through network striping and more efficient data paths). The principal contribution of this paper is to demonstrate the Together, these advantages influence a large enough fraction of the potential of network-attached storage devices for penetrating the storage market to make commodity network-attached storage fea- markets defined by existing distributed file system clients, specifi- sible. Realizing the technology’s full potential requires careful cally the Network File System (NFS) and Andrew File System consideration across a wide range of file system, networking and (AFS) distributed file system protocols. -
An Advanced Solution for Long-Term Retention of Enterprise Digital
Data Management, Simplified. Enterprise Edition An Advanced Solution for Long-term Retention of Enterprise Digital Assets A centrally managed, secure archive is a key requirement for today’s enterprises, which must contend with exponential growth in data volumes, long-term retention requirements, and e-discovery mandates. Atempo Digital Archive, Enterprise Edition, is a comprehensive solution that addresses each of these data archiving challenges. Providing capabilities for both automatically and manually archiving data in long-term storage platforms, Atempo Digital Archive ensures that all corporate data is easy to find and access for as long as it needs to be retained. An Answer to Each Archiving Need Addressing Today’s Storage Management Challenges Enabling the Long-term Retention of Digital Assets Today, organizations must manage more, and more Digital information represents both the lifeblood of a critical, data than ever before—and data volumes continue business today, and the basis of records that may need to grow at an explosive pace. Traditional backup and to be accessed for generations. With its integrated recovery solutions have proven ill-equipped to keep up capabilities for full content indexing, search, and metadata with these demands. To ensure these increasing volumes indexing, Atempo Digital Archive allows users to retrieve of critical information are always available when needed, information quickly and easily. In addition, it offers organizations require advanced new archiving capabilities. refreshing mechanisms that verify whether storage Atempo Digital Archive is the one solution that enables media is still readable, and that ensure information IT organizations to address their storage management can continue to be accessed on an ongoing basis. -
Advfs Command Line and Application Programming Interface
AdvFS Command Line and Application Programming Interface External Reference Specification Version 1.13 JA CASL Not Inspected/Date Inspected Building ZK3 110 Spit Brook Road Nashua, NH 03062 Copyright (C) 2008 Hewlett-Packard Development Company, L.P. Page 1 of 65 Preface Version 1.4 of the AdvFS Command Line and API External Reference Specification is being made available to all partners in order to allow them to design and implement code meeting the specifications contained herein. 1 INTRODUCTION ....................................................................................................................................................................4 1.1 ABSTRACT .........................................................................................................................................................................4 1.2 AUDIENCE ..........................................................................................................................................................................4 1.3 TERMS AND DEFINITIONS ...................................................................................................................................................4 1.4 RELATED DOCUMENTS ......................................................................................................................................................4 1.5 ACKNOWLEDGEMENTS ......................................................................................................................................................4 -
Overview Addressing Today's Storage Management Challenges Enabling
DATA PROTECTION ATEMPO-DIGITAL ARCHIVE High performance file archiving software for large data volumes OVERVIEW A centrally managed, secure archive is a key requirement for today’s enterprises, which must contend with exponential growth in data volumes and long-term retention requirements. Atempo-Digital Archive is a comprehensive solution that addresses each of these data archiving challenges. Providing capabilities for both automatically and manually archiving large data volumes in long-term storage platforms, Atempo-Digital Archive ensures that all corporate data is easy to find and access for as long as it needs to be retained. ADDRESSING TODAY’S STORAGE MANAGEMENT CHALLENGES Today, organizations must manage more critical data than ever before - and data volumes continue to grow at an explosive pace. Traditional backup and recovery solutions have proven ill-equipped to keep up with these demands. To ensure these increasing volumes of critical information are always available when needed, organizations require advanced new archiving capabilities. Atempo-Digital Archive (ADA) is the one solution that enables IT organizations to address their storage management challenges today and in the long term. By managing data in a secure, centrally-managed archive, Atempo-Digital Archive represents the most efficient, reliable, and cost-effective way to protect critical business information. This solution ensures that an organization’s most valuable data is organized and easily accessible - while at the same time freeing up expensive disk space, lowering storage costs, and reducing the backup window for data that is infrequently accessed. ENABLING THE LONG-TERM RETENTION OF DIGITAL ASSETS Digital information represents both the lifeblood of a business today, and the basis of records that may need to be accessed for generations. -
Petascale Data Management: Guided by Measurement
Petascale Data Management: Guided by Measurement petascale data storage institute www.pdsi-scidac.org/ MPP2 www.pdsi-scidac.org MEMBER ORGANIZATIONS & Lustre • Los Alamos National Laboratory – institute.lanl.gov/pdsi/ • Parallel Data Lab, Carnegie Mellon University – www.pdl.cmu.edu/ • Oak Ridge National Laboratory – www.csm.ornl.gov/ • Sandia National Laboratories – www.sandia.gov/ • National Energy Research Scientific Computing Center • Center for Information Technology Integration, U. of Michigan pdsi.nersc.gov/ www.citi.umich.edu/projects/pdsi/ • Pacific Northwest National Laboratory – www.pnl.gov/ • University of California at Santa Cruz – www.pdsi.ucsc.edu/ The Computer Failure Data Repository Filesystems Statistics Survey • Goal: to collect and make available failure data from a large variety of sites GOALS • Better understanding of the characteristics of failures in the real world • Gather & build large DB of static filetree summary • Now maintained by USENIX at cfdr.usenix.org/ • Build small, non-invasive, anonymizing stats gather tool • Distribute fsstats tool via easily used web site Red Storm NAME SYSTEM TYPE SYSTEM SIZE TIME PERIOD TYPE OF DATA • Encourage contributions (output of tool) from many FSs Any node • Offer uploaded statistics & summaries to public & Lustre 22 HPC clusters 5000 nodes 9 years outage . Label Date Type File Total Size Total Space # files # dirs max size max space max dir max name avg file avg dir . 765 nodes (2008) System TB TB M K GB GB ents bytes MB ents . 1 HPC cluster 5 years PITTSBURGH 3,400 disks -
The Influence of Scale on Distributed File System Design
IEEE TRANSAmIONS ON SOFIWARE ENGINEERING, VOL. 18, NO. I, JANUARY lY92 The Influence of Scale on Distributed File System Design Mahadev Satyanarayanan, Member, IEEE Abstract- Scale should be recognized as a primary factor into autonomous or semi-autonomous organizations for man- influencing the architecture and implementation of distributed agement purposes. Hence a distributed system that has many systems. This paper uses Andrew and Coda, distributed file users or nodes will also span many organizations. Regardless systems built at Carnegie Mellon University, to validate this proposition. Performance, operability, and security are dominant of the specific metric of scale, the designs of distributed considerations in the design of these systems. Availability is a systems that scale well are fundamentally different from less further consideration in the design of Coda. Client caching, scalable designs. bulk data transfer, token-based mutual authentication, and hi- In this paper we describe the lessons we have learned erarchical organization of the protection domain have emerged about scalability from the Andrew File System and the Codu as mechanisms that enhance scalability. The separation of con- cerns made possible by functional specialization has also proved File System. These systems are particularly appropriate for valuable in scaling. Heterogeneity is an important by-product our discussion, because they have been designed for growth of growth, but the mechanisms available to cope with it are to thousands of nodes and users. We focus on distributed rudimentary. Physical separation of clients and servers turns out file systems, because they are the most widely used kind of to be a critical requirement for scalability. -
Distributed File Systems
Please note: Please start working your research project (proposal will be due on Feb. 19 in class) Each group needs to turn in a printed version of their proposal and intermediate report. Also, before class each group needs to email me a DOC version. Instructor’s Guide for Coulouris, Dollimore and Kindberg Distributed Systems: Concepts and Design Edn. 4 © Pearson Education 2005 Remote Procedure Call (1): at-least-once or at-most-once semantics client: "stub" instead of "proxy" (same function, different names) behaves like a local procedure, marshal arguments, communicate the request server: dispatcher "stub": unmarshal arguments, communicate the results back Instructor’s Guide for Coulouris, Dollimore and Kindberg Distributed Systems: Concepts and Design Edn. 4 © Pearson Education 2005 Remote Procedure Call (2) client process server process Request Reply client stub server stub procedure procedure client service program Communication Communication procedure module module dispatcher Instructor’s Guide for Coulouris, Dollimore and Kindberg Distributed Systems: Concepts and Design Edn. 4 © Pearson Education 2005 Sun RPC (1): Designed for client-server communication in the SUN NFS (network file system) Supplied as a part of SUN and other UNIX operating systems Over either UDP or TCP Provides an interface definition language (IDL) initially XDR is for data representation, extended to be IDL less modern than CORBA IDL and Java program numbers (obtained from a central authority) instead of interface names procedure numbers (used as a procedure identifier) instead of procedure names only a single input parameter is allowed (then we have to use a ?) Offers an interface compiler (rpcgen) for C language, which generates the following: client stub server main procedure, dispatcher, and server stub XDR marshalling, unmarshaling Instructor’s Guide for Coulouris, Dollimore and Kindberg Distributed Systems: Concepts and Design Edn.