An Introduction to NFS

Total Page:16

File Type:pdf, Size:1020Kb

An Introduction to NFS An Introduction to NFS Avishay Traeger IBM Haifa Research Lab Internal Storage Course ― November 2010 v1.2 Outline The Basics NFSv2 NFSv3 NFSv4 NFSv4.1 2 Typical Use ws-avishay ws-bob ws-carl mount -t nfs (NFS Client) (NFS Client) (NFS Client) nfsserv:/home /home 10.0.2.56 10.0.2.103 10.0.2.81 nfsserv (NFS Server) /etc/exports: / /home 10.0.2.*(rw) home … avishay bob carl Some benefits of NFS: 1. All clients have the same view 2. Centralized storage management RAID Storage 3 NFS Evolution NFS is a standardized protocol Version Year RFC # Pages Status NFSv2 1989 1094 27 Obsolete NFSv3 1995 1813 126 Most popular Available on several OSs, slowly NFSv4 2003 3530 275 but surely replacing NFSv3 NFSv4.1 2010 5661 617 Early adopters only 4 Design Goals OS independence & interoperability Simple crash recovery for clients and servers Transparent access (client programs do not know files are remote) Maintain local file system semantics Reasonable performance 5 Remote Procedure Call (RPC) NFS is defined as a set of RPCs – their arguments, results, and effects RPCs are synchronous The use of RPCs makes the protocol easier to understand 6 NFS Client/Server 7 Outline The Basics NFSv2 NFSv3 NFSv4 NFSv4.1 8 Stateless Protocol The server does not keep state for RPCs Each RPC contains the necessary information to complete the call This makes crash recovery easy Server crash: server does no crash recovery, clients resubmit requests Client crash: no crash recovery for client or server This is nice in theory, but Adds complexity Not really stateless... File locking adds state, provided by separate protocol & daemon Server keeps an RPC reply cache to handle duplicate non- idempotent RPC 9 File Handles The most common NFS procedure parameter is a structure called a file handle (fh, fhandle) Provided by the server and used by the client to reference a file The fhandle is opaque to the client New fhandles returned by LOOKUP, CREATE, MKDIR, ... The fhandle for the root of the file system is obtained by the client when it mounts the file system 10 Operations NULL() returns () Do nothing procedure to used for pinging the server LOOKUP(dirfh, name) returns (fh, attr) Returns a new fh and attributes for the named file in the directory specified by dirfh CREATE(dirfh, name, attr) returns (newfh, attr) Creates a new file name in the directory dirfh and returns the new fh and attributes. REMOVE(dirfh, name) returns (status) Removes the file name from directory dirfh. 11 Operations GETATTR(fh) returns (attr) Returns file attributes (similar to stat syscall) SETATTR(fh, attr) returns (attr) Sets the mode, uid, gid, size, access time, and modify time of a file. Setting the size to zero truncates the file. 12 Operations READ(fh, offset, count) returns (attr, data) Returns up to count bytes of data from a file starting offset bytes into the file. Returns the attributes of the file. WRITE(fh, offset, count, data) returns (attr) Writes count bytes of data to a file beginning at offset bytes from the beginning of the file. Returns the new attributes of the file after the write. 13 Operations RENAME(dirfh, name, tofh, toname) returns (status) Renames name in directory dirfh, to toname in directory tofh. LINK(dirfh, name, tofh, toname) returns (status) Creates a hard link toname in directory tofh, that points to name in directory dirfh. 14 Operations SYMLINK(dirfh, name, string) returns (status) Creates a symlink name in the directory dirfh with value string. The server does not interpret the string argument in any way, just saves it and makes an association to the new symlink file. READLINK(fh) returns (string) Returns the string which is associated with the symlink file. 15 Operations MKDIR(dirfh, name, attr) returns (fh, newattr) Creates a new directory name in the directory dirfh and returns the new fh and attributes. RMDIR(dirfh, name) returns (status) Removes the empty directory name from the parent directory dirfh. STATFS(fh) returns (fsstats) Returns file system information such as block size, number of free blocks, etc. 16 Operations READDIR (dirfh, cookie, count) returns (entries) Returns up to count bytes of directory entries from the directory dirfh. Each entry contains a file name, file id, and an opaque pointer to the next directory entry called a cookie. The cookie is used in subsequent readdir calls to start reading at a specific entry in the directory. A readdir call with the cookie of zero returns entries starting with the first entry in the directory. 17 The MOUNT Protocol The MOUNT protocol takes a directory pathname and returns an fhandle if the client has permissions to mount the file system Separate protocol Easier to plug in new permission check methods Separates the OS-dependent aspects of the protocol Other OS implementations can change the MOUNT protocol without having to change the NFS protocol 18 The Linux File Handle Remember that information contained in the fhandle is only meaningful on the server If the local FS on the server reuses an inode number, an NFS client could mistakenly use an old file handle and access the new file. File systems include generation numbers in the inode to avoid this. The value is usually taken from a counter used across the file system. Important file handle fields: Major/minor number of the exported device Inode number Generation number 19 Security NFSv2 uses UNIX-style permission checks The client passes uid/gid info in RPCs, and the server performs permission checks as if the user was performing the operation locally Problem – the mapping from uid/gid to user must be the same on the client and server Can be solved via Network Information Service (NIS) Another problem – should root on the client have root access to files on the server? Server specifies policy 20 Cache Consistency Problems Clients use caching and write buffering to improve performance, but this causes issues Problem: Update visibility; If client C1 buffers writes in its cache, client C2 will see the old version NFSv2 solution: Close-to-open consistency – Clients flush on close(), so other clients will see the latest version on open() Problem: Stale cache; If C1 has a file cached, it will see old data even if the file is updated by C2 NFSv2 solution: Send a GETATTR and check the file's modification time to see if it has been updated. Cache attributes for a few seconds to reduce the number of GETATTR calls. 21 Strong Semantics for Write Because the NFS server is stateless, when servicing an NFS request it must commit any modified data to stable storage before returning results The implication for UNIX based servers is that requests which modify the file system must flush all modified data & metadata to disk before returning from the call This can be a big performance bottleneck unless something is done to improve write performance (e.g., NetApp's WAFL file system) 22 Outline The Basics NFSv2 NFSv3 NFSv4 NFSv4.1 23 Major Changes from NFSv2 to v3 Sizes and offsets are widened from 32 bits to 64 bits A new COMMIT RPC allows for reliable asynchronous writes A new ACCESS RPC improves support for ACLs and super-user All operations now return attributes to reduce the number of subsequent GETATTR procedure calls The 8KB data size limitation on the READ and WRITE procedures is relaxed 24 Major Changes from NFSv2 to v3 A new READDIRPLUS RPC returns both file handle and attributes to eliminate LOOKUP calls when scanning a directory 25 Asynchronous Writes In NFSv3, the server can reply to WRITE RPCs immediately, without syncing to disk When the client wants to ensure that data is on stable storage, it sends a COMMIT RPC Asynchronous writes are optional, and negotiated at mount time 26 Asynchronous Writes: Crash Recovery The client must keep all uncommitted data in case of a server crash Replies for WRITE and COMMIT RPCs include a write verifier for server crash detection Write verifier: 8-byte value that the server must change if it crashes (many use boot time) The client must save verifiers returned by async WRITE RPCs, and compare them to the verifier returned by a leter COMMIT RPC If the verifiers don't match, the client must rewrite all uncommitted data 27 Outline The Basics NFSv2 NFSv3 NFSv4 NFSv4.1 28 Additional Goals for NFSv4 Improved access and good performance on the Internet Only TCP Easy to transit firewalls: uses one port (mount & lock protocols merged into NFS) COMPOUNDs, delegations, uid/gid issue resolved Strong security with negotiation built in Better cross-platform interoperability Better extensibility New security types, new attributes, etc. Big design change – NFSv4 is stateful 29 Security For previous versions, only UNIX permissions were widely adopted NFSv4 mandates the use of strong RPC security flavors that depend on cryptography Security type negotiation is done securely and in-band User and groups are identified with strings, not numbers Access control policies compatible with both UNIX and Windows The problematic MOUNT protocol is removed 30 RPCSEC_GSS A framework adopted by NFSv4 to provide authentication, integrity, and privacy at the RPC level The following mechanisms must be implemented: Kerberos v5, LIPKEY, SPKM3 Security options are negotiated at mount time The SECINFO operation allows a client to determine the security policy (usually on mount, but can be on a per-filehandle basis) RPCSEC_GSS can be used with previous versions of NFS, but in NFSv4 support is mandatory 31 Identifying
Recommended publications
  • CST8207 – Linux O/S I
    Mounting a Filesystem Directory Structure Fstab Mount command CST8207 - Algonquin College 2 Chapter 12: page 467 - 496 CST8207 - Algonquin College 3 The mount utility connects filesystems to the Linux directory hierarchy. The mount point is a directory in the local filesystem where you can access mounted filesystem. This directory must exist before you can mount a filesystem. All filesystems visible on the system exist as a mounted filesystem someplace below the root (/) directory CST8207 - Algonquin College 4 can be mounted manually ◦ can be listed in /etc/fstab, but not necessary ◦ all mounting information supplied manually at command line by user or administrator can be mounted automatically on startup ◦ must be listed /etc/fstab, with all appropriate information and options required Every filesystem, drive, storage device is listed as a mounted filesystem associated to a directory someplace under the root (/) directory CST8207 - Algonquin College 5 CST8207 - Algonquin College 6 Benefits Scalable ◦ As new drives are added and new partitions are created, further filesystems can be mounted at various mount points as required. ◦ This means a Linux system does not need to worry about running out of disk space. Transparent ◦ No application would stop working if transferred to a different partition, because access to data is done via the mount point. ◦ Also transparent to user CST8207 - Algonquin College 7 All known filesystems volumes are typically listed in the /etc/fstab (static information about filesystem) file to help automate the mounting process If it is not listed in the /etc/fstab file, then all appropriate information about the filesystem needs to be listed manually at the command line.
    [Show full text]
  • STAT 625: Statistical Case Studies 1 Our Computing Environment(S)
    John W. Emerson, Department of Statistics, Yale University © 2013 1 STAT 625: Statistical Case Studies John W. Emerson Yale University Abstract This term, I’ll generally present brief class notes and scripts, but much of the lecture material will simply be presented in class and/or available online. However, student participation is a large part of this course, and at any time I may call upon any of you to share some of your work. Be prepared. This won’t be a class where you sit and absorb material. I strongly encourage you to work with other students; you can learn a lot from each other. To get anything out of this course you’ll need to put in the time for the sake of learning and strengthening your skills, not for the sake of jumping through the proverbial hoop. Lesson 1: Learning how to communicate. This document was created using LATEX and Sweave (which is really an R package). There are some huge advantages to a workflow like this, relating to a topic called reproducible research. Here’s my thought: pay your dues now, reap the benefits later. Don’t trust me on everything, but trust me on this. A sister document was created using R Markdown and the markdown package, and this is also acceptable. Microsoft Office is not. So, in today’s class I’ll show you how I produced these documents and I’ll give you the required files. I recommend you figure out how to use LATEX on your own computer. You already have Sweave, because it comes with R.
    [Show full text]
  • System Calls and I/O
    System Calls and I/O CS 241 January 27, 2012 Copyright ©: University of Illinois CS 241 Staff 1 This lecture Goals Get you familiar with necessary basic system & I/O calls to do programming Things covered in this lecture Basic file system calls I/O calls Signals Note: we will come back later to discuss the above things at the concept level Copyright ©: University of Illinois CS 241 Staff 2 System Calls versus Function Calls? Copyright ©: University of Illinois CS 241 Staff 3 System Calls versus Function Calls Function Call Process fnCall() Caller and callee are in the same Process - Same user - Same “domain of trust” Copyright ©: University of Illinois CS 241 Staff 4 System Calls versus Function Calls Function Call System Call Process Process fnCall() sysCall() OS Caller and callee are in the same Process - Same user - OS is trusted; user is not. - Same “domain of trust” - OS has super-privileges; user does not - Must take measures to prevent abuse Copyright ©: University of Illinois CS 241 Staff 5 System Calls System Calls A request to the operating system to perform some activity System calls are expensive The system needs to perform many things before executing a system call The computer (hardware) saves its state The OS code takes control of the CPU, privileges are updated. The OS examines the call parameters The OS performs the requested function The OS saves its state (and call results) The OS returns control of the CPU to the caller Copyright ©: University of Illinois CS 241 Staff 6 Steps for Making a System Call
    [Show full text]
  • File System, Files, and *Tab /Etc/Fstab
    File system, files, and *tab File system files directories volumes, file systems mounting points local versus networked file systems 1 /etc/fstab Specifies what is to be mounted where and how fs_spec: describes block special device for remote filesystem to be mounted fs_file: describes the mount point fs_vfstype: describes the type of file system fs_mntops: describes the mount options associated with the filesystem 2 /etc/fstab cont. fs_freq: used by the dump command fs_passno: used by fsck to determine the order in which checks are done at boot time. Root file systems should be specified as 1, others should be 2. Value 0 means that file system does not need to be checked 3 /etc/fstab 4 from blocks to mounting points metadata inodes directories superblocks 5 mounting file systems mounting e.g., mount -a unmounting manually or during shutdown umount 6 /etc/mtab see what is mounted 7 Network File System Access file system (FS) over a network looks like a local file system to user e.g. mount user FS rather than duplicating it (which would be a disaster) Developed by Sun Microsystems (mid 80s) history for NFS: NFS, NFSv2, NFSv3, NFSv4 RFC 3530 (from 2003) take a look to see what these RFCs are like!) 8 Network File System How does this actually work? server needs to export the system client needs to mount the system server: /etc/exports file client: /etc/fstab file 9 Network File System Security concerns UID GID What problems could arise? 10 Network File System example from our raid system (what is a RAID again?) Example of exports file from
    [Show full text]
  • Lab #6: Stat Overview Stat
    February 17, 2017 CSC 357 1 Lab #6: Stat Overview The purpose of this lab is for you to retrieve from the filesystem information pertaining to specified files. You will use portions of this lab in Assignment #4, so be sure to write code that can be reused. stat Write a simplified version of the stat utility available on some Unix distributions. This program should take any number of file names as command-line arguments and output information pertaining to each file as described below (you should use the stat system call). The output from your program should look as follows: File: ‘Makefile’ Type: Regular File Size: 214 Inode: 4702722 Links: 1 Access: -rw------- The stat system call provides more information about a file, but this lab only requires the above. The format of each field follows: • File: the name of the file being examined. • Type: the type of file output as “Regular File”, “Directory”, “Character Device”, “Block Device”, “FIFO”, or “Symbolic Link” • Size: the number of bytes in the file. • Inode: the file’s inode. • Links: the number of links to the file. • Access: the access permissions for the file. This field is the most involved since the output string can take many forms. The first bit of the access permissions denotes the type (similar to the second field above). The valid characters for the first bit are: b – Block special file. c – Character special file. d – Directory. l – Symbolic link. p – FIFO. - – Regular file. The remaining bits denote the read, write and execute permissions for the user, group, and others.
    [Show full text]
  • Filesystem Considerations for Embedded Devices ELC2015 03/25/15
    Filesystem considerations for embedded devices ELC2015 03/25/15 Tristan Lelong Senior embedded software engineer Filesystem considerations ABSTRACT The goal of this presentation is to answer a question asked by several customers: which filesystem should you use within your embedded design’s eMMC/SDCard? These storage devices use a standard block interface, compatible with traditional filesystems, but constraints are not those of desktop PC environments. EXT2/3/4, BTRFS, F2FS are the first of many solutions which come to mind, but how do they all compare? Typical queries include performance, longevity, tools availability, support, and power loss robustness. This presentation will not dive into implementation details but will instead summarize provided answers with the help of various figures and meaningful test results. 2 TABLE OF CONTENTS 1. Introduction 2. Block devices 3. Available filesystems 4. Performances 5. Tools 6. Reliability 7. Conclusion Filesystem considerations ABOUT THE AUTHOR • Tristan Lelong • Embedded software engineer @ Adeneo Embedded • French, living in the Pacific northwest • Embedded software, free software, and Linux kernel enthusiast. 4 Introduction Filesystem considerations Introduction INTRODUCTION More and more embedded designs rely on smart memory chips rather than bare NAND or NOR. This presentation will start by describing: • Some context to help understand the differences between NAND and MMC • Some typical requirements found in embedded devices designs • Potential filesystems to use on MMC devices 6 Filesystem considerations Introduction INTRODUCTION Focus will then move to block filesystems. How they are supported, what feature do they advertise. To help understand how they compare, we will present some benchmarks and comparisons regarding: • Tools • Reliability • Performances 7 Block devices Filesystem considerations Block devices MMC, EMMC, SD CARD Vocabulary: • MMC: MultiMediaCard is a memory card unveiled in 1997 by SanDisk and Siemens based on NAND flash memory.
    [Show full text]
  • Mv-Ch650-90Tm
    MV-CH650-90TM 65 MP CMOS 10 GigE Area Scan Camera Introduction Available Model MV-CH650-90TM camera adopts Gpixel GMAX3265 sensor to M58-mount with fan, mono: MV-CH650- provide high-quality image. It uses 10 GigE interface to transmit 90TM-M58S-NF non-compressed image in real time, and its max. frame rate can F-mount with fan, mono: MV-CH650-90TM- reach 15.5 fps in full resolution. F-NF Key Feature Applicable Industry Resolution of 9344 × 7000, and pixel size of 3.2 μm × 3.2 μm. PCB AOI, FPD, railway related applications, etc. Adopts 10 GigE interface providing max. transmission Sensor Quantum Efficiency distance of 100 meters without relay. Supports auto or manual adjustment for gain, exposure time, and manual adjustment for Look-Up Table (LUT), Gamma correction, etc. Compatible with GigE Vision Protocol V2.0, GenlCam Standard, and third-party software based on protocols. Dimension M58-mount with fan: F-mount with fan: Specification Model MV-CH650-90TM Camera Sensor type CMOS, global shutter Sensor model Gpixel GMAX3265 Pixel size 3.2 µm × 3.2 µm Sensor size 29.9 mm × 22.4 mm Resolution 9344 × 7000 Max. frame rate 15.5 fps @9344 × 7000 Dynamic range 66 dB SNR 40 dB Gain 1.25X to 6X Exposure time 15 μs to 10 sec Exposure mode Off/Once/Continuous exposure mode Mono/color Mono Pixel format Mono 8/10/10p/12/12p Binning Supports 1 × 1, 1 × 2, 1 × 4, 2 × 1, 2 × 2, 2 × 4, 4 × 1, 4 × 2, 4 × 4 Decimation Supports 1 × 1, 1 × 2, 1 × 4, 2 × 1, 2 × 2, 2 × 4, 4 × 1, 4 × 2, 4 × 4 Reverse image Supports horizontal and vertical reverse image output Electrical features Data interface 10 Gigabit Ethernet, compatible with Gigabit Ethernet Digital I/O 12-pin Hirose connector provides power and I/O, including opto-isolated input × 1 (Line 0), opto-isolated output × 1 (Line 1), bi-directional non-isolated I/O × 1 (Line 2), and RS-232 × 1 Power supply 9 VDC to 24 VDC Power consumption Typ.
    [Show full text]
  • How to Setup NFS File System Guide ID: 3 - Release: Initial Revision [Major] 2015-08-14
    How to setup NFS file system Guide ID: 3 - Release: Initial revision [major] 2015-08-14 How to setup NFS file system Configuration of SCO Unix shared drive in order to share printer tasks. Written By: Petr Roupec This document was generated on 2020-11-19 05:38:51 AM (MST). © 2020 omlex.dozuki.com/ Page 1 of 7 How to setup NFS file system Guide ID: 3 - Release: Initial revision [major] 2015-08-14 INTRODUCTION This guide is describing use of SCO scoadmin program to setup mount /volumes/bmprint remote drive on your OT computer. This document was generated on 2020-11-19 05:38:51 AM (MST). © 2020 omlex.dozuki.com/ Page 2 of 7 How to setup NFS file system Guide ID: 3 - Release: Initial revision [major] 2015-08-14 Step 1 — SCO Admin - Starting program Switch user on local console. Please note character "-" on command line - this load right environment for root user From remote computer use telnet connection Start scoadmin program This document was generated on 2020-11-19 05:38:51 AM (MST). © 2020 omlex.dozuki.com/ Page 3 of 7 How to setup NFS file system Guide ID: 3 - Release: Initial revision [major] 2015-08-14 Step 2 — SCO Admin - File System Manager SCO Admin - File Manager Select FileSystems Open Filesystem Manager Use TAB and arrows on your keyboard to move between the fields This document was generated on 2020-11-19 05:38:51 AM (MST). © 2020 omlex.dozuki.com/ Page 4 of 7 How to setup NFS file system Guide ID: 3 - Release: Initial revision [major] 2015-08-14 Step 3 — Start NFS file system mounting wizard Select Mount from menu Scroll down and select "Add Mount Configuration" Choose remote Use TAB and arrows on your keyboard to move between the fields Step 4 — NFS Share - Configuration details Enter IP address of your printer server Enter name of remote directory of your print server Enter name of local directory on computer you are configuring Don't forget Advanced Mount Option - Failure to configure these correctly might stop your server in case of printer server shutdown This document was generated on 2020-11-19 05:38:51 AM (MST).
    [Show full text]
  • UNIX (Solaris/Linux) Quick Reference Card Logging in Directory Commands at the Login: Prompt, Enter Your Username
    UNIX (Solaris/Linux) QUICK REFERENCE CARD Logging In Directory Commands At the Login: prompt, enter your username. At the Password: prompt, enter ls Lists files in current directory your system password. Linux is case-sensitive, so enter upper and lower case ls -l Long listing of files letters as required for your username, password and commands. ls -a List all files, including hidden files ls -lat Long listing of all files sorted by last Exiting or Logging Out modification time. ls wcp List all files matching the wildcard Enter logout and press <Enter> or type <Ctrl>-D. pattern Changing your Password ls dn List files in the directory dn tree List files in tree format Type passwd at the command prompt. Type in your old password, then your new cd dn Change current directory to dn password, then re-enter your new password for verification. If the new password cd pub Changes to subdirectory “pub” is verified, your password will be changed. Many systems age passwords; this cd .. Changes to next higher level directory forces users to change their passwords at predetermined intervals. (previous directory) cd / Changes to the root directory Changing your MS Network Password cd Changes to the users home directory cd /usr/xx Changes to the subdirectory “xx” in the Some servers maintain a second password exclusively for use with Microsoft windows directory “usr” networking, allowing you to mount your home directory as a Network Drive. mkdir dn Makes a new directory named dn Type smbpasswd at the command prompt. Type in your old SMB passwword, rmdir dn Removes the directory dn (the then your new password, then re-enter your new password for verification.
    [Show full text]
  • TS ODBC Dataserver Quick Start
    TS ODBC DataServerTM Quick Start Multiple-Tier Introduction This Multiple-Tier product includes 3 components. Follow the instructions below for each Windows workstation and DataServer Host component. Multiple-Tier components can be found by platform in a folder on the TS ODBC DataServer CD-ROM. Use these instructions for the TS ODBC Gateway for Windows version of the Multiple-Tier software. TS ODBC DataServer Server UNIX Server Install the Server on your UNIX Host system from cpio distribution media. This installation is required only once no matter how many workstations are connected. Logon as root. 1. Create and change (cd) to a base directory for the TS ODBC DataServer (For example, /usr/local/tsodbc). 2. Copy the distribution media to the system using cpio. (See Mounting UNIX CD-ROM devices on the reverse.) This example is for Linux (kernel 2.6.16+). Substitute the appropriate values for your environment. umask 0 cpio -icvBmud </mountpoint/linux2616/tsod_srv/tsod (for Linux use –ivBmud above) 3. Execute the install script. ./install 4. Activate the server (Refer to the Installation and Activation Guide). Windows Server Before continuing, review the updated installation instructions provided in the installation manual. NOTE: All Thoroughbred Windows based products prior to Version 8.7.0 must first be uninstalled and then the 8.7.1 release installed. Only 8.7.0 can be upgraded to 8.7.1 and only 8.7.0 and 8.7.1 can co-exist on the same system. If you are upgrading a pre 8.7.0 release, BEFORE continuing with this installation, please see the TS ODBC Installation and Activation Guide for complete instructions to properly prepare your system for 8.7.1.
    [Show full text]
  • HAWK MV-4000 Configuration Guide Ethernet Configuration with Flying Leads
    HAWK MV-4000 Smart Camera Configuration Guide Hardware Required Item Description Part Number 1 HAWK MV-4000 Smart Camera 8X1X-XXX0-010X 2 Lens, C-Mount 98-90001XX-01 3 IP67 Lens Cover for HAWK MV-4000, 50 mm or 70 mm (not shown) 98-900015X-01 4 Lens Extension Tube Set, 0.5 mm, 1 mm, 5 mm, 10 mm, 20 mm, 40 mm 98-CO206 5 Cable, HAWK MV-4000 Ethernet, X-CODE / RJ45 CAT 6A, 1 m, 3 m, or 5 m 61-9000134-0X 6 Cable, HAWK MV-4000 M12 to USB Socket or VGA / USB, 1 m 61-900014X-01 7 Cable, Adapter, HAWK MV-4000 to Accessory Cables / Power Supply (supplied with camera) 61-9000132-01 8 Cable, HAWK MV-4000 M12 to Flying Leads, 3 m (no adapter required) 61-9000151-01 9 QX Cordset, HAWK MV-4000 Adapter to QX-1 M12 Plug (Screw-On), 1 m or 3 m 61-0001XX-02 10 QX-1 Interface Device 98-000103-02 QX Photo Sensor, M12 4-Pin Plug, NPN, Dark On or Dark Off, 2 m or 99-000020-0X 11 Trigger Connector, 4-Pin Plug (Screw Terminal, Field-Wireable) (Self-Wiring) 20-610024-01 12 Y Cable, HAWK MV-4000 Adapter to Smart Series Illuminator and QX-1, Power or On/Off or Strobe, 1 m 61-900013X-01 13 Cable, QX-1 to Smart Series Illuminator, Continuous Power or On/Off or Strobe 61-0002XX-01 14 Power Supply, 100-240VAC, +24VDC, M12 12-Pin Socket 97-000012-01 15 Universal Mount, HAWK MV-4000 98-9000209-01 Note: See page 5 of this document for a full list of available accessories, numbered to correspond with this table and the diagrams below.
    [Show full text]
  • Exercise University of Oklahoma, May 13Th – 17Th 2019 J.D
    Linux Clusters Institute: Lustre Hands On Exercise University of Oklahoma, May 13th – 17th 2019 J.D. Maloney | Storage Engineer National Center for Supercomputing Applications (NCSA) [email protected] This document is a result of work by volunteer LCI instructors and is licensed under CC BY-NC 4.0 (https://creativecommons.org/licenses/by-nc/4.0/). Goal of Hands on Exercise • Create Lustre File System with 1 MDT & 6 OSTs • Bring in Sample Data • Configure Quotas • Test how stripe width works May 13th-17th 2019 2 Lay of the Land • You should have 4 storage servers; 1 for metadata, 3 for data • Each have 2 5GB volumes • Root SSH between all nodes in cluster • Chronyd keeping time in sync May 13th-17th 2019 3 Creating Lustre File System • Add Lustre Repos to Servers & Compute Nodes (create /etc/yum.repos.d/lustre.repo) [lustre-server] name=CentOS-$releasever - Lustre baseurl=https://downloads.hpdd.intel.com/public/lustre/latest-feature-release/el7/server/ gpgcheck=0 [e2fsprogs] name=CentOS-$releasever - Ldiskfs baseurl=https://downloads.hpdd.intel.com/public/e2fsprogs/latest/el7/ gpgcheck=0 [lustre-client] name=CentOS-$releasever - Lustre baseurl=https://downloads.hpdd.intel.com/public/lustre/latest-feature-release/el7/client/ gpgcheck=0 May 13th-17th 2019 4 Creating Lustre File System • Update/Install these packages on the four storage nodes yum upgrade -y e2fsprogs yum install -y lustre-tests • Create /etc/modprobe.d/lnet.conf and add the following on all four storage nodes options lnet networks=tcp0(eth0) • Reboot all four storage
    [Show full text]