SUGI 23: It's Only Temporary
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Better Performance Through a Disk/Persistent-RAM Hybrid Design
The Conquest File System: Better Performance Through a Disk/Persistent-RAM Hybrid Design AN-I ANDY WANG Florida State University GEOFF KUENNING Harvey Mudd College PETER REIHER, GERALD POPEK University of California, Los Angeles ________________________________________________________________________ Modern file systems assume the use of disk, a system-wide performance bottleneck for over a decade. Current disk caching and RAM file systems either impose high overhead to access memory content or fail to provide mechanisms to achieve data persistence across reboots. The Conquest file system is based on the observation that memory is becoming inexpensive, which enables all file system services to be delivered from memory, except providing large storage capacity. Unlike caching, Conquest uses memory with battery backup as persistent storage, and provides specialized and separate data paths to memory and disk. Therefore, the memory data path contains no disk-related complexity. The disk data path consists of only optimizations for the specialized disk usage pattern. Compared to a memory-based file system, Conquest incurs little performance overhead. Compared to several disk-based file systems, Conquest achieves 1.3x to 19x faster memory performance, and 1.4x to 2.0x faster performance when exercising both memory and disk. Conquest realizes most of the benefits of persistent RAM at a fraction of the cost of a RAM-only solution. Conquest also demonstrates that disk-related optimizations impose high overheads for accessing memory content in a memory-rich environment. Categories and Subject Descriptors: D.4.2 [Operating Systems]: Storage Management—Storage Hierarchies; D.4.3 [Operating Systems]: File System Management—Access Methods and Directory Structures; D.4.8 [Operating Systems]: Performance—Measurements General Terms: Design, Experimentation, Measurement, and Performance Additional Key Words and Phrases: Persistent RAM, File Systems, Storage Management, and Performance Measurement ________________________________________________________________________ 1. -
System Calls System Calls
System calls We will investigate several issues related to system calls. Read chapter 12 of the book Linux system call categories file management process management error handling note that these categories are loosely defined and much is behind included, e.g. communication. Why? 1 System calls File management system call hierarchy you may not see some topics as part of “file management”, e.g., sockets 2 System calls Process management system call hierarchy 3 System calls Error handling hierarchy 4 Error Handling Anything can fail! System calls are no exception Try to read a file that does not exist! Error number: errno every process contains a global variable errno errno is set to 0 when process is created when error occurs errno is set to a specific code associated with the error cause trying to open file that does not exist sets errno to 2 5 Error Handling error constants are defined in errno.h here are the first few of errno.h on OS X 10.6.4 #define EPERM 1 /* Operation not permitted */ #define ENOENT 2 /* No such file or directory */ #define ESRCH 3 /* No such process */ #define EINTR 4 /* Interrupted system call */ #define EIO 5 /* Input/output error */ #define ENXIO 6 /* Device not configured */ #define E2BIG 7 /* Argument list too long */ #define ENOEXEC 8 /* Exec format error */ #define EBADF 9 /* Bad file descriptor */ #define ECHILD 10 /* No child processes */ #define EDEADLK 11 /* Resource deadlock avoided */ 6 Error Handling common mistake for displaying errno from Linux errno man page: 7 Error Handling Description of the perror () system call. -
Serverless Network File Systems
Serverless Network File Systems Thomas E. Anderson, Michael D. Dahlin, Jeanna M. Neefe, David A. Patterson, Drew S. Roselli, and Randolph Y. Wang Computer Science Division University of California at Berkeley Abstract In this paper, we propose a new paradigm for network file system design, serverless network file systems. While traditional network file systems rely on a central server machine, a serverless system utilizes workstations cooperating as peers to provide all file system services. Any machine in the system can store, cache, or control any block of data. Our approach uses this location independence, in combination with fast local area networks, to provide better performance and scalability than traditional file systems. Further, because any machine in the system can assume the responsibilities of a failed component, our serverless design also provides high availability via redundant data storage. To demonstrate our approach, we have implemented a prototype serverless network file system called xFS. Preliminary performance measurements suggest that our architecture achieves its goal of scalability. For instance, in a 32-node xFS system with 32 active clients, each client receives nearly as much read or write throughput as it would see if it were the only active client. 1. Introduction A serverless network file system distributes storage, cache, and control over cooperating workstations. This approach contrasts with traditional file systems such as Netware [Majo94], NFS [Sand85], Andrew [Howa88], and Sprite [Nels88] where a central server machine stores all data and satisfies all client cache misses. Such a central server is both a performance and reliability bottleneck. A serverless system, on the other hand, distributes control processing and data storage to achieve scalable high performance, migrates the responsibilities of failed components to the remaining machines to provide high availability, and scales gracefully to simplify system management. -
NASD) Storage Architecture, Prototype Implementations of [Patterson88]
Proceedings of the 8th Conference on Architectural Support for Programming Languages and Operating Systems, 1998. A Cost-Effective, High-Bandwidth Storage Architecture Garth A. Gibson*, David F. Nagle†, Khalil Amiri†, Jeff Butler†, Fay W. Chang*, Howard Gobioff*, Charles Hardin†, Erik Riedel†, David Rochberg*, Jim Zelenka* School of Computer Science* Department of Electrical and Computer Engineering† Carnegie Mellon University, Pittsburgh, PA 15213 [email protected] ABSTRACT width, that is, linearly increasing application bandwidth with increasing numbers of storage devices and client processors, This paper describes the Network-Attached Secure Disk the data must be striped over many disks and network links (NASD) storage architecture, prototype implementations of [Patterson88]. With 1998 technology, most office, engineer- NASD drives, array management for our architecture, and ing, and data processing shops have sufficient numbers of three filesystems built on our prototype. NASD provides scal- disks and scalable switched networking, but they access stor- able storage bandwidth without the cost of servers used age through storage controller and distributed fileserver primarily for transferring data from peripheral networks bottlenecks. These bottlenecks arise because a single (e.g. SCSI) to client networks (e.g. ethernet). Increasing “server” computer receives data from the storage (periph- eral) network and forwards it to the client (local area) dataset sizes, new attachment technologies, the convergence network while adding -
File Manager Manual
FileManager Operations Guide for Unisys MCP Systems Release 9.069W November 2017 Copyright This document is protected by Federal Copyright Law. It may not be reproduced, transcribed, copied, or duplicated by any means to or from any media, magnetic or otherwise without the express written permission of DYNAMIC SOLUTIONS INTERNATIONAL, INC. It is believed that the information contained in this manual is accurate and reliable, and much care has been taken in its preparation. However, no responsibility, financial or otherwise, can be accepted for any consequence arising out of the use of this material. THERE ARE NO WARRANTIES WHICH EXTEND BEYOND THE PROGRAM SPECIFICATION. Correspondence regarding this document should be addressed to: Dynamic Solutions International, Inc. Product Development Group 373 Inverness Parkway Suite 110, Englewood, Colorado 80112 (800)641-5215 or (303)754-2000 Technical Support Hot-Line (800)332-9020 E-Mail: [email protected] ii November 2017 Contents ................................................................................................................................ OVERVIEW .......................................................................................................... 1 FILEMANAGER CONSIDERATIONS................................................................... 3 FileManager File Tracking ................................................................................................ 3 File Recovery .................................................................................................................... -
File Permissions Do Not Restrict Root
Filesystem Security 1 General Principles • Files and folders are managed • A file handle provides an by the operating system opaque identifier for a • Applications, including shells, file/folder access files through an API • File operations • Access control entry (ACE) – Open file: returns file handle – Allow/deny a certain type of – Read/write/execute file access to a file/folder by – Close file: invalidates file user/group handle • Access control list (ACL) • Hierarchical file organization – Collection of ACEs for a – Tree (Windows) file/folder – DAG (Linux) 2 Discretionary Access Control (DAC) • Users can protect what they own – The owner may grant access to others – The owner may define the type of access (read/write/execute) given to others • DAC is the standard model used in operating systems • Mandatory Access Control (MAC) – Alternative model not covered in this lecture – Multiple levels of security for users and documents – Read down and write up principles 3 Closed vs. Open Policy Closed policy Open Policy – Also called “default secure” • Deny Tom read access to “foo” • Give Tom read access to “foo” • Deny Bob r/w access to “bar” • Give Bob r/w access to “bar • Tom: I would like to read “foo” • Tom: I would like to read “foo” – Access denied – Access allowed • Tom: I would like to read “bar” • Tom: I would like to read “bar” – Access allowed – Access denied 4 Closed Policy with Negative Authorizations and Deny Priority • Give Tom r/w access to “bar” • Deny Tom write access to “bar” • Tom: I would like to read “bar” – Access -
Data File Manager
Data File Manager Welcome to Data File Manager! Thank you for using the data file manager. Highlights of the data file manager are as follows: Intuitive operability Buttons on the tool menu, drag & drop, and context menu with a right click are supported to use each function. Tree view for data in SD memory card Files and folders in SD memory card are shown in a tree structure and target files can easily be found. Operations for files and folders Files and folders can be added, renamed or deleted. PLC maintenance Viewing the PLC status, such as Run/Stop state or error status, downloading/uploading user programs, or upgrading the system software of the PLC are supported. MICRO/I maintenance Viewing the target information, such as system software version or external memory device information, or downloading/uploading project is supported. Notes - Data File Manager There are the following notes when you use Data File Manager: Supported hardware Data File Manager supports the following hardware: FC4A series MICROSmart FC5A series MICROSmart Pentra FC6A series MICROSmart FT1A series SmartAXIS expect FT1A Touch HG3G/4G series MICRO/I HG2G-5F series MICRO/I HG2G-5T series MICRO/I HG1G series MICRO/I SD card access (SmartAXIS Pro & Lite) SD card inserted on SmartAXIS series can be accessed while the PLC is stopped only. Downloading or uploading the user program (PLC) You can download or upload the user programs using ZLD files, which can be created or read with WindLDR. Downloading or uploading the project (MICRO/I) You can download or upload the project using ZNV files, which can be created or read with WindO/I-NV4. -
Life in the Fast Lane: Optimisation for Interactive and Batch Jobs Nikola Marković, Boemska T.S
Paper 11825-2016 Life in the Fast Lane: Optimisation for Interactive and Batch Jobs Nikola Marković, Boemska T.S. Ltd.; Greg Nelson, ThotWave Technologies LLC. ABSTRACT We spend so much time talking about GRID environments, distributed jobs, and huge data volumes that we ignore the thousands of relatively tiny programs scheduled to run every night, which produce only a few small data sets, but make all the difference to the users who depend on them. Individually, these jobs might place a negligible load on the system, but by their sheer number they can often account for a substantial share of overall resources, sometimes even impacting the performance of the bigger, more important jobs. SAS® programs, by their varied nature, use available resources in a varied way. Depending on whether a SAS® procedure is CPU-, disk- or memory-intensive, chunks of memory or CPU can sometimes remain unused for quite a while. Bigger jobs leave bigger chunks, and this is where being small and able to effectively exploit those leftovers can be a great advantage. We call our main optimization technique the Fast Lane, which is a queue configuration dedicated to jobs with consistently small SASWORK directories, that, when available, lets them use unused RAM in place of their SASWORK disk. The approach improves overall CPU saturation considerably while taking loads off the I/O subsystem, and without failure results in improved runtimes for big jobs and small jobs alike, without requiring any changes to deployed code. This paper explores the practical aspects of implementing the Fast Lane on your environment. -
File Server Scaling with Network-Attached Secure Disks
Appears in Proceedings of the ACM International Conference on Measurement and Modeling of Computer Systems (Sigmetrics ‘97), Seattle, Washington, June 15-18, 1997. File Server Scaling with Network-Attached Secure Disks Garth A. Gibson†, David F. Nagle*, Khalil Amiri*, Fay W. Chang†, Eugene M. Feinberg*, Howard Gobioff†, Chen Lee†, Berend Ozceri*, Erik Riedel*, David Rochberg†, Jim Zelenka† *Department of Electrical and Computer Engineering †School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213-3890 [email protected] http://www.cs.cmu.edu/Web/Groups/NASD/ Abstract clients with efficient, scalable, high-bandwidth access to stored data. This paper discusses a powerful approach to fulfilling this By providing direct data transfer between storage and client, net- need. Network-attached storage provides high bandwidth by work-attached storage devices have the potential to improve scal- directly attaching storage to the network, avoiding file server ability for existing distributed file systems (by removing the server store-and-forward operations and allowing data transfers to be as a bottleneck) and bandwidth for new parallel and distributed file striped over storage and switched-network links. systems (through network striping and more efficient data paths). The principal contribution of this paper is to demonstrate the Together, these advantages influence a large enough fraction of the potential of network-attached storage devices for penetrating the storage market to make commodity network-attached storage fea- markets defined by existing distributed file system clients, specifi- sible. Realizing the technology’s full potential requires careful cally the Network File System (NFS) and Andrew File System consideration across a wide range of file system, networking and (AFS) distributed file system protocols. -
Winreporter Documentation
WinReporter documentation Table Of Contents 1.1 WinReporter overview ............................................................................................... 1 1.1.1.................................................................................................................................... 1 1.1.2 Web site support................................................................................................. 1 1.2 Requirements.............................................................................................................. 1 1.3 License ....................................................................................................................... 1 2.1 Welcome to WinReporter........................................................................................... 2 2.2 Scan requirements ...................................................................................................... 2 2.3 Simplified Wizard ...................................................................................................... 3 2.3.1 Simplified Wizard .............................................................................................. 3 2.3.2 Computer selection............................................................................................. 3 2.3.3 Validate .............................................................................................................. 4 2.4 Advanced Wizard....................................................................................................... 5 2.4.1 -
The Zebra Striped Network File System
The Zebra Striped Network File System John H. Hartman John K. Ousterhout Computer Science Division Electrical Engineering and Computer Sciences University of California Berkeley, CA 94720 Abstract speed of its processor, network interface, I/O busses, and disks. It is possible to split a file system among multiple Zebra is a network file system that increases throughput servers but each file must reside on a single server and it is by striping file data across multiple servers. Rather than difficult to balance the loads of the different servers. For striping each file separately, Zebra forms all the new data example, the system directories often lie on a single server, from each client into a single stream, which it then stripes making that server a hot spot. using an approach similar to a log-structured file system. In the future, new styles of computing such as multi- This provides high performance for writes of small files as media and parallel computation are likely to demand much well as for reads and writes of large files. Zebra also writes greater throughput than today’s applications, making the parity information in each stripe in the style of RAID disk limitations of a single server even more severe. For arrays; this increases storage costs slightly but allows the example, a single video playback can consume a substantial fraction of a file server’s bandwidth even when the video is system to continue operation even while a single storage compressed. A cluster of workstations can easily exceed the server is unavailable. A prototype implementation of Zebra, bandwidth of a file server if they all run video applications built in the Sprite operating system, provides 4-5 times the simultaneously, and the problems will become much worse throughput of the standard Sprite file system or NFS for when video resolution increases with the arrival of HDTV. -
007-2007: the FILENAME Statement Revisited
SAS Global Forum 2007 Applications Development Paper 007-2007 The FILENAME Statement Revisited Yves DeGuire, Statistics Canada, Ottawa, Ontario, Canada ABSTRACT The FILENAME statement has been around for a long time for connecting external files with SAS®. Over the years, it has grown quite a bit to encompass capabilities that go beyond the simple sequential file. Most SAS programmers have used the FILENAME statement to read/write external files stored on disk. However, few programmers have used it to access devices other than disk or to interface with different access methods. This paper focuses on some of those capabilities by showing you some interesting devices and access methods supported by the FILENAME statement. INTRODUCTION – FILENAME STATEMENT 101 The FILENAME statement is a declarative statement that is used to associate a fileref with a physical file or a device. A fileref is a logical name that can be used subsequently in lieu of a physical name. The syntax for the FILENAME statement is as follows: FILENAME fileref <device-type> <'external-file'> <options>; This is the basic syntax to associate an external file or a device to a fileref. There are other forms that allow you to list or clear filerefs. Please refer to SAS OnlineDoc® documentation for a complete syntax of the FILENAME statement. Once the association has been made between a physical file and a fileref, the fileref can be used in as many DATA steps as necessary without having to re-run a FILENAME statement. Within a DATA step, the use of an INFILE or a FILE statement allows the programmer to determine which file to read or write using an INPUT or a PUT statement.