Manual Pages (Man) for OCFS2 1.8 Release

Total Page:16

File Type:pdf, Size:1020Kb

Manual Pages (Man) for OCFS2 1.8 Release OCFS2(7) OCFS2 Manual Pages OCFS2(7) NAME OCFS2 − A Shared-Disk Cluster File System for Linux INTRODUCTION OCFS2 is a file system.Itallows users to store and retrieve data. The data is stored in files that are orga- nized in a hierarchical directory tree. It is a POSIX compliant file system that supports the standard inter- faces and the behavioral semantics as spelled out by that specification. It is also a shared disk cluster file system, one that allows multiple nodes to access the same disk at the same time. This is where the fun begins as allowing a file system to be accessible on multiple nodes opens a can of worms. What if the nodes are of different architectures? What if a node dies while writing to the file system? What data consistencycan one expect if processes on twonodes are reading and writing concur- rently? What if one node removesafile while it is still being used on another node? Unlikemost shared file systems where the answer is fuzzy,the answer in OCFS2 is very well defined. It behavesonall nodes exactly likealocal file system. If a file is removed, the directory entry is removedbut the inode is kept as long as it is in use across the cluster.When the last user closes the descriptor,the inode is marked for deletion. The data consistencymodel follows the same principle. It works as if the twoprocesses that are running on twodifferent nodes are running on the same node. A read on a node gets the last write irrespective ofthe IO mode used. The modes can be buffered, direct, asynchronous, splice or memory mapped IOs. It is fully cache coherent. Take for example the REFLINK feature that allows a user to create multiple write-able snapshots of a file. This feature, likeall others, is fully cluster-aware. A file being written to on multiple nodes can be safely reflinked on another.The snapshot created is a point-in-time image of the file that includes both the file data and all its attributes (including extended attributes). It is a journaling file system. When a node dies, a surviving node transparently replays the journal of the dead node. This ensures that the file system metadata is always consistent. It also defaults to ordered data journaling to ensure the file data is flushed to disk before the journal commit, to remove the small possibil- ity of stale data appearing in files after a crash. It is architecture and endian neutral.Itallows concurrent mounts on nodes with different processors like x86, x86_64, IA64 and PPC64. It handles little and big endian, 32-bit and 64-bit architectures. It is featurerich.Itsupports indexed directories, metadata checksums, extended attributes, POSIX ACLs, quotas, REFLINKs, sparse files, unwritten extents and inline-data. It is fully integrated with the mainline Linux kernel. The file system was merged into Linux kernel 2.6.16 in early 2006. It is quickly installed.Itisavailable with almost all Linux distributions. The file system is on-disk com- patible across all of them. It is modular.The file system can be configured to operate with other cluster stacks like Pacemaker and CMAN along with its own stack, O2CB. It is easily configured.The O2CB cluster stack configuration involves editing twofiles, one for cluster lay- out and the other for cluster timeouts. It is very efficient.The file system consumes very little resources. It is used to store virtual machine images in limited memory environments likeXen and KVM. Version 1.8.2 January 2012 1 OCFS2(7) OCFS2 Manual Pages OCFS2(7) In summary,OCFS2 is an efficient, easily configured, modular,quickly installed, fully integrated and com- patible, feature-rich, architecture and endian neutral, cache coherent, ordered data journaling, POSIX-com- pliant, shared disk cluster file system. OVERVIEW OCFS2 is a general-purpose shared-disk cluster file system for Linux capable of providing both high per- formance and high availability. As it provides local file system semantics, it can be used with almost all applications. Cluster-aware appli- cations can makeuse of cache-coherent parallel I/Os from multiple nodes to scale out applications easily. Other applications can makeuse of the clustering facilities to fail-overrunning application in the event of a node failure. The notable features of the file system are: Tunable Block size The file system supports block sizes of 512, 1K, 2K and 4K bytes. 4KB is almost always recom- mended. This feature is available in all releases of the file system. Tunable Cluster size Acluster size is also referred to as an allocation unit. The file system supports cluster sizes of 4K, 8K, 16K, 32K, 64K, 128K, 256K, 512K and 1M bytes. For most use cases, 4KB is recommended. However, a larger value is recommended for volumes hosting mostly very large files likedatabase files, virtual machine images, etc. A large cluster size allows the file system to store large files more efficiently.This feature is available in all releases of the file system. Endian and Architectureneutral The file system can be mounted concurrently on nodes having different architectures. Like32-bit, 64-bit, little-endian (x86, x86_64, ia64) and big-endian (ppc64, s390x). This feature is available in all releases of the file system. Buffered, Direct, Asynchronous, Splice and Memory Mapped I/O modes The file system supports all modes of I/O for maximum flexibility and performance. It also sup- ports cluster-wide shared writeable mmap(2).The support for bufferred, direct and asynchronous I/O is available in all releases. The support for splice I/O was added in Linux kernel 2.6.20 and for shared writeable map(2) in 2.6.23. Multiple Cluster Stacks The file system includes a flexible framework to allowittofunction with userspace cluster stacks likePacemaker (pcmk)and CMAN (cman), its own in-kernel cluster stack o2cb and no cluster stack. The support for o2cb cluster stack is available in all releases. The support for no cluster stack, or local mount, was added in Linux kernel 2.6.20. The support for userspace cluster stack was added in Linux kernel 2.6.26. Journaling The file system supports both ordered (default) and writeback data journaling modes to provide file system consistencyinthe event of power failure or system crash. It uses JBD2 in Linux kernel 2.6.28 and later.Itused JBD in earlier kernels. Version 1.8.2 January 2012 2 OCFS2(7) OCFS2 Manual Pages OCFS2(7) Extent-based Allocations The file system allocates and tracks space in ranges of clusters. This is unlikeblock based file sys- tems that have totrack each and every block. This feature allows the file system to be very effi- cient when dealing with both large volumes and large files. This feature is available in all releases of the file system. Sparse files Sparse files are files with holes. With this feature, the file system delays allocating space until a write is issued to a cluster.This feature was added in Linux kernel 2.6.22 and requires enabling on-disk feature sparse. Unwritten Extents An unwritten extent is also referred to as user pre-allocation. It allows an application to request a range of clusters to be allocated, but not initialized, within a file. Pre-allocation allows the file sys- tem to optimize the data layout with fewer,larger extents. It also provides a performance boost, delaying initialization until the user writes to the clusters. This feature was added in Linux kernel 2.6.23 and requires enabling on-disk feature unwritten. Hole Punching Hole punching allows an application to remove arbitrary allocated regions within a file. Creating holes, essentially.This is more efficient than zeroing the same extents. This feature is especially useful in virtualized environments as it allows a block discard in a guest file system to be con- verted to a hole punch in the host file system thus allowing users to reduce disk space usage. This feature was added in Linux kernel 2.6.23 and requires enabling on-disk features sparse and unwritten. Inline-data Inline data is also referred to as data-in-inode as it allows storing small files and directories in the inode block. This not only savesspace but also has a positive impact on cold-cache directory and file operations. The data is transparently movedout to an extent when it no longer fits inside the inode block. This feature was added in Linux kernel 2.6.24 and requires enabling on-disk feature inline-data. REFLINK REFLINK is also referred to as fast copy. Itallows users to atomically (and instantly) copyregular files. In other words, create multiple writeable snapshots of regular files. It is called REFLINK because it looks and feels more likea(hard) link(2) than a traditional snapshot. Likea link, it is a regular user operation, subject to the security attributes of the inode being reflinked and not to the super user privileges typically required to create a snapshot. Likealink, it operates within a file system. But unlikealink, it links the inodes at the data extent levelallowing each reflinked inode to growindependently as and when written to. Up to four billion inodes can share a data extent. This feature was added in Linux kernel 2.6.32 and requires enabling on-disk feature refcount. Allocation Reservation File contiguity plays an important role in file system performance. When a file is fragmented on disk, reading and writing to the file involves manyseeks, leading to lower throughput. Contiguous files, on the other hand, minimize seeks, allowing the disks to perform IO at the maximum rate. With allocation reservation, the file system reserves a windowinthe bitmap for all extending files allowing each to growascontiguously as possible.
Recommended publications
  • A Brief History of UNIX File Systems
    A Brief History of UNIX File Systems Val Henson IBM, Inc. [email protected] Summary • Review of UNIX file system concepts • File system formats, 1974-2004 • File system comparisons and recommendations • Fun trivia • Questions and answers (corrections ONLY during talk) 1 VFS/vnode architecture • VFS: Virtual File System: common object-oriented interface to fs's • vnode: virtual node: abstract file object, includes vnode ops • All operations to fs's and files done through VFS/vnode in- terface • S.R. Kleiman, \Vnodes: An Architecture for Multiple File System Types in Sun UNIX," Summer USENIX 1986 2 Some Definitions superblock: fs summary, pointers to other information inode: on-disk structure containing information about a file indirect block: block containing pointers to other blocks metadata: everything that is not user data, including directory entries 3 Disk characteristics • Track - contiguous region, can be read at maximum speed • Seek time - time to move the head between different tracks • Rotational delay - time for part of track to move under head • Fixed per I/O overhead means bigger I/Os are better 4 In the beginning: System V FS (S5FS) (c. 1974) • First UNIX file system, referred to as \FS" • Disk layout: superblock, inodes, followed by everything else • 512-1024 byte block size, no fragments • Super simple - and super slow! 2-5% of raw disk bandwidth 5 Berkeley Fast File System (FFS or UFS) (c. 1984) • Metadata spread throughout the disk in \cylinder groups" • Block size 4KB minimum, frag size 1KB (to avoid 45% wasted space) • Physical
    [Show full text]
  • The Linux Command Line
    The Linux Command Line Fifth Internet Edition William Shotts A LinuxCommand.org Book Copyright ©2008-2019, William E. Shotts, Jr. This work is licensed under the Creative Commons Attribution-Noncommercial-No De- rivative Works 3.0 United States License. To view a copy of this license, visit the link above or send a letter to Creative Commons, PO Box 1866, Mountain View, CA 94042. A version of this book is also available in printed form, published by No Starch Press. Copies may be purchased wherever fine books are sold. No Starch Press also offers elec- tronic formats for popular e-readers. They can be reached at: https://www.nostarch.com. Linux® is the registered trademark of Linus Torvalds. All other trademarks belong to their respective owners. This book is part of the LinuxCommand.org project, a site for Linux education and advo- cacy devoted to helping users of legacy operating systems migrate into the future. You may contact the LinuxCommand.org project at http://linuxcommand.org. Release History Version Date Description 19.01A January 28, 2019 Fifth Internet Edition (Corrected TOC) 19.01 January 17, 2019 Fifth Internet Edition. 17.10 October 19, 2017 Fourth Internet Edition. 16.07 July 28, 2016 Third Internet Edition. 13.07 July 6, 2013 Second Internet Edition. 09.12 December 14, 2009 First Internet Edition. Table of Contents Introduction....................................................................................................xvi Why Use the Command Line?......................................................................................xvi
    [Show full text]
  • Filesystem Hierarchy Standard
    Filesystem Hierarchy Standard LSB Workgroup, The Linux Foundation Filesystem Hierarchy Standard LSB Workgroup, The Linux Foundation Version 3.0 Publication date March 19, 2015 Copyright © 2015 The Linux Foundation Copyright © 1994-2004 Daniel Quinlan Copyright © 2001-2004 Paul 'Rusty' Russell Copyright © 2003-2004 Christopher Yeoh Abstract This standard consists of a set of requirements and guidelines for file and directory placement under UNIX-like operating systems. The guidelines are intended to support interoperability of applications, system administration tools, development tools, and scripts as well as greater uniformity of documentation for these systems. All trademarks and copyrights are owned by their owners, unless specifically noted otherwise. Use of a term in this document should not be regarded as affecting the validity of any trademark or service mark. Permission is granted to make and distribute verbatim copies of this standard provided the copyright and this permission notice are preserved on all copies. Permission is granted to copy and distribute modified versions of this standard under the conditions for verbatim copying, provided also that the title page is labeled as modified including a reference to the original standard, provided that information on retrieving the original standard is included, and provided that the entire resulting derived work is distributed under the terms of a permission notice identical to this one. Permission is granted to copy and distribute translations of this standard into another language, under the above conditions for modified versions, except that this permission notice may be stated in a translation approved by the copyright holder. Dedication This release is dedicated to the memory of Christopher Yeoh, a long-time friend and colleague, and one of the original editors of the FHS.
    [Show full text]
  • Set Hadoop-Specific Environment Variables Here
    # Set Hadoop-specific environment variables here. # The only required environment variable is JAVA_HOME. All others are # optional. When running a distributed configuration it is best to # set JAVA_HOME in this file, so that it is correctly defined on # remote nodes. # The java implementation to use. Required. export JAVA_HOME=/usr/jdk64/jdk1.8.0_112 export HADOOP_HOME_WARN_SUPPRESS=1 # Hadoop home directory export HADOOP_HOME=${HADOOP_HOME:-/usr/hdp/2.6.5.0-292/hadoop} # Hadoop Configuration Directory # Path to jsvc required by secure HDP 2.0 datanode export JSVC_HOME=/usr/lib/bigtop-utils # The maximum amount of heap to use, in MB. Default is 1000. export HADOOP_HEAPSIZE="1024" export HADOOP_NAMENODE_INIT_HEAPSIZE="-Xms1024m" # Extra Java runtime options. Empty by default. export HADOOP_OPTS="-Djava.net.preferIPv4Stack=true ${HADOOP_OPTS}" USER="$(whoami)" # Command specific options appended to HADOOP_OPTS when specified HADOOP_JOBTRACKER_OPTS="-server -XX:ParallelGCThreads=8 -XX:+UseConcMarkSweepGC - XX:ErrorFile=/var/log/hadoop/$USER/hs_err_pid%p.log -XX:NewSize=200m -XX:MaxNewSize=200m - Xloggc:/var/log/hadoop/$USER/gc.log-`date +'%Y%m%d%H%M'` -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps - XX:+PrintGCDateStamps -Xmx1024m -Dhadoop.security.logger=INFO,DRFAS -Dmapred.audit.logger=INFO,MRAUDIT - Dhadoop.mapreduce.jobsummary.logger=INFO,JSA ${HADOOP_JOBTRACKER_OPTS}" HADOOP_TASKTRACKER_OPTS="-server -Xmx1024m -Dhadoop.security.logger=ERROR,console -Dmapred.audit.logger=ERROR,console ${HADOOP_TASKTRACKER_OPTS}" SHARED_HADOOP_NAMENODE_OPTS="-server
    [Show full text]
  • Oracle® Linux 7 Managing File Systems
    Oracle® Linux 7 Managing File Systems F32760-07 August 2021 Oracle Legal Notices Copyright © 2020, 2021, Oracle and/or its affiliates. This software and related documentation are provided under a license agreement containing restrictions on use and disclosure and are protected by intellectual property laws. Except as expressly permitted in your license agreement or allowed by law, you may not use, copy, reproduce, translate, broadcast, modify, license, transmit, distribute, exhibit, perform, publish, or display any part, in any form, or by any means. Reverse engineering, disassembly, or decompilation of this software, unless required by law for interoperability, is prohibited. The information contained herein is subject to change without notice and is not warranted to be error-free. If you find any errors, please report them to us in writing. If this is software or related documentation that is delivered to the U.S. Government or anyone licensing it on behalf of the U.S. Government, then the following notice is applicable: U.S. GOVERNMENT END USERS: Oracle programs (including any operating system, integrated software, any programs embedded, installed or activated on delivered hardware, and modifications of such programs) and Oracle computer documentation or other Oracle data delivered to or accessed by U.S. Government end users are "commercial computer software" or "commercial computer software documentation" pursuant to the applicable Federal Acquisition Regulation and agency-specific supplemental regulations. As such, the use, reproduction, duplication, release, display, disclosure, modification, preparation of derivative works, and/or adaptation of i) Oracle programs (including any operating system, integrated software, any programs embedded, installed or activated on delivered hardware, and modifications of such programs), ii) Oracle computer documentation and/or iii) other Oracle data, is subject to the rights and limitations specified in the license contained in the applicable contract.
    [Show full text]
  • Filesystem Considerations for Embedded Devices ELC2015 03/25/15
    Filesystem considerations for embedded devices ELC2015 03/25/15 Tristan Lelong Senior embedded software engineer Filesystem considerations ABSTRACT The goal of this presentation is to answer a question asked by several customers: which filesystem should you use within your embedded design’s eMMC/SDCard? These storage devices use a standard block interface, compatible with traditional filesystems, but constraints are not those of desktop PC environments. EXT2/3/4, BTRFS, F2FS are the first of many solutions which come to mind, but how do they all compare? Typical queries include performance, longevity, tools availability, support, and power loss robustness. This presentation will not dive into implementation details but will instead summarize provided answers with the help of various figures and meaningful test results. 2 TABLE OF CONTENTS 1. Introduction 2. Block devices 3. Available filesystems 4. Performances 5. Tools 6. Reliability 7. Conclusion Filesystem considerations ABOUT THE AUTHOR • Tristan Lelong • Embedded software engineer @ Adeneo Embedded • French, living in the Pacific northwest • Embedded software, free software, and Linux kernel enthusiast. 4 Introduction Filesystem considerations Introduction INTRODUCTION More and more embedded designs rely on smart memory chips rather than bare NAND or NOR. This presentation will start by describing: • Some context to help understand the differences between NAND and MMC • Some typical requirements found in embedded devices designs • Potential filesystems to use on MMC devices 6 Filesystem considerations Introduction INTRODUCTION Focus will then move to block filesystems. How they are supported, what feature do they advertise. To help understand how they compare, we will present some benchmarks and comparisons regarding: • Tools • Reliability • Performances 7 Block devices Filesystem considerations Block devices MMC, EMMC, SD CARD Vocabulary: • MMC: MultiMediaCard is a memory card unveiled in 1997 by SanDisk and Siemens based on NAND flash memory.
    [Show full text]
  • Linux System Administration
    Linux System Administration Jonathan Quick Hartebeesthoek Radio Astronomy Observatory Goals • Help you to understand how Linux starts up, keeps running, and shuts down • Give confidence in dealing with hardware and software failures • Give an overview of what you can configure and how • Show you where to find more information when you need it • For the field system and Mark5’s 2 Basic Linux Concepts • Linux Kernel – Base monolithic kernel + loadable modules – Gives standardized access to underlying hardware • Linux System / "Distribution" – Kernel + lots of software – Adds both system and application level software to the system • Background processes ("daemons") 3 System Modifications • In order to do any system-wide changes you usually have to be logged in as 'root‘ – Or have root privileges • There are a number of approaches for this – Log in as user “root” – Execute “su –” from the present user account – Execute the command directly with “sudo” • E.g. “sudo tail /var/log/kern.log” 4 Logging in as 'root' • You can change to a virtual console (Ctrl-Alt- F1) and login normally or use 'su -' • 'root' can override all permissions, start and stop anything, erase hard drives,... – So please be careful with disk names and similar! – You can browse and check many (if not most of the) things as a normal user (like 'oper'). 5 Sudo • Sudo is a program designed to allow a sysadmin to give limited root privileges to users and log root activity. • The basic philosophy is to give as few privileges as possible but still allow people to get their work
    [Show full text]
  • Parallel File Systems for HPC Introduction to Lustre
    Parallel File Systems for HPC Introduction to Lustre Piero Calucci Scuola Internazionale Superiore di Studi Avanzati Trieste November 2008 Advanced School in High Performance and Grid Computing Outline 1 The Need for Shared Storage 2 The Lustre File System 3 Other Parallel File Systems Parallel File Systems for HPC Cluster & Storage Piero Calucci Shared Storage Lustre Other Parallel File Systems A typical cluster setup with a Master node, several computing nodes and shared storage. Nodes have little or no local storage. Parallel File Systems for HPC Cluster & Storage Piero Calucci The Old-Style Solution Shared Storage Lustre Other Parallel File Systems • a single storage server quickly becomes a bottleneck • if the cluster grows in time (quite typical for initially small installations) storage requirements also grow, sometimes at a higher rate • adding space (disk) is usually easy • adding speed (both bandwidth and IOpS) is hard and usually involves expensive upgrade of existing hardware • e.g. you start with an NFS box with a certain amount of disk space, memory and processor power, then add disks to the same box Parallel File Systems for HPC Cluster & Storage Piero Calucci The Old-Style Solution /2 Shared Storage Lustre • e.g. you start with an NFS box with a certain amount of Other Parallel disk space, memory and processor power File Systems • adding space is just a matter of plugging in some more disks, ot ar worst adding a new controller with an external port to connect external disks • but unless you planned for
    [Show full text]
  • Altavault Cloud Integrated Storage Command-Line Reference Guide
    NetApp® AltaVault® Cloud Integrated Storage 4.2.1 Command-Line Reference Guide NetApp, Inc. Telephone: +1 (408) 822-6000 Part number: 215-11346_A0 495 East Java Drive Fax: + 1 (408) 822-4501 August 2016 Sunnyvale, CA 94089 Support telephone: +1(888) 463-8277 U.S. Web: www.netapp.com Feedback: [email protected] Contents Beta Draft Contents Chapter 1 - Using the Command-Line Interface ......................................................................................3 Connecting to the CLI ......................................................................................................................................3 Overview of the CLI.........................................................................................................................................4 Entering Commands .........................................................................................................................................5 Accessing CLI Online Help..............................................................................................................................5 Error Messages .................................................................................................................................................5 Command Negation..........................................................................................................................................5 Running the Configuration Wizard...................................................................................................................6
    [Show full text]
  • Ubuntu Server Guide Basic Installation Preparing to Install
    Ubuntu Server Guide Welcome to the Ubuntu Server Guide! This site includes information on using Ubuntu Server for the latest LTS release, Ubuntu 20.04 LTS (Focal Fossa). For an offline version as well as versions for previous releases see below. Improving the Documentation If you find any errors or have suggestions for improvements to pages, please use the link at thebottomof each topic titled: “Help improve this document in the forum.” This link will take you to the Server Discourse forum for the specific page you are viewing. There you can share your comments or let us know aboutbugs with any page. PDFs and Previous Releases Below are links to the previous Ubuntu Server release server guides as well as an offline copy of the current version of this site: Ubuntu 20.04 LTS (Focal Fossa): PDF Ubuntu 18.04 LTS (Bionic Beaver): Web and PDF Ubuntu 16.04 LTS (Xenial Xerus): Web and PDF Support There are a couple of different ways that the Ubuntu Server edition is supported: commercial support and community support. The main commercial support (and development funding) is available from Canonical, Ltd. They supply reasonably- priced support contracts on a per desktop or per-server basis. For more information see the Ubuntu Advantage page. Community support is also provided by dedicated individuals and companies that wish to make Ubuntu the best distribution possible. Support is provided through multiple mailing lists, IRC channels, forums, blogs, wikis, etc. The large amount of information available can be overwhelming, but a good search engine query can usually provide an answer to your questions.
    [Show full text]
  • Isolation File Systems
    Physical Disentanglement in a Container-Based File System Lanyue Lu, Yupu Zhang, Thanh Do, Samer AI-Kiswany Andrea C. Arpaci-Dusseau, Remzi H. Arpaci-Dusseau University of Wisconsin Madison Motivation Isolation in various subsystems ➡ resources: virtual machines, Linux containers ➡ security: BSD jail, sandbox ➡ reliability: address spaces File systems lack isolation ➡ physical entanglement in modern file systems Three problems due to entanglement ➡ global failures, slow recovery, bundled performance Global Failures Definition ➡ a failure which impacts all users of the file system or even the operating system Read-Only ➡ mark the file system as read-only ➡ e.g., metadata corruption, I/O failure Crash ➡ crash the file system or the operating system ➡ e.g., unexpected states, pointer faults Ext3 200 Read-Only Crash 150 100 50 Number of Failure Instances 0 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Linux 3.X Versions Ext4 Read-Only Crash 400 350 300 250 200 150 100 50 Number of Failure Instances 0 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Linux 3.X Versions Slow Recovery File system checker ➡ repair a corrupted file system ➡ usually offline Current checkers are not scalable ➡ increasing disk capacities and file system sizes ➡ scan the whole file system ➡ checking time is unacceptably long Scalability of fsck on Ext3 Ext3 1007 1000 800 723 600 476 400 Fsck Time (s) 231 200 0 200GB 400GB 600GB 800GB File-system Capacity Bundled Performance Shared transaction ➡ all updates share a single transaction ➡ unrelated workloads affect each other Consistency guarantee
    [Show full text]
  • Linux System Administration
    Linux System Administration Jonathan Quick, Hartebeesthoek Radio Astronomy Observatory Ari Mujunen, Metsähovi Radio Observatory Linux Startup and Shutdown Managing Hard Disks, Partitions, Backups Rescuing a Failing PC / System Modifying Configuration Adding/Removing Packages 1 Goals Help you to understand how Linux starts up, keeps running, and shuts down Give confidence in dealing with hardware and software failures Give an overview of what you can configure and how Show you where to find more information when you need it 2 Basic Linux Concepts Linux Kernel Base monolithic kernel + loadable modules Gives standardized access to underlying hardware Linux System / "Distribution" Kernel + lots of software Adds both system and application level software to the system Background processes ("daemons") 3 Logging in as 'root' In order to do any system-wide changes you usually have to be logged in as 'root' You can change to a virtual console (Ctrl-Alt- F1) and login normally or use 'su -' 'root' can override all permissions, start and stop anything, erase hard drives,... So please be careful with disk names and similar! You can browse and check many (if not most of the) things as a normal user (like 'oper'). 4 Getting System Information ps axf, top; kill, kill -9 free df, mount netstat -an, ifconfig, route -n w, who cat /proc/cpuinfo (and others) 5 Linux PC-Level Startup PC ROM BIOS initializes hardware and boots a Master Boot Record (MBR) From a floppy, hard disk, CD-ROM, ... That MBR contains LILO, the Linux Loader
    [Show full text]