Freebsd Enterprise Storage Polish BSD User Group Welcome 2020/02/11 Freebsd Enterprise Storage

Total Page:16

File Type:pdf, Size:1020Kb

Freebsd Enterprise Storage Polish BSD User Group Welcome 2020/02/11 Freebsd Enterprise Storage FreeBSD Enterprise Storage Polish BSD User Group Welcome 2020/02/11 FreeBSD Enterprise Storage Sławomir Wojciech Wojtczak [email protected] vermaden.wordpress.com twitter.com/vermaden bsd.network/@vermaden https://is.gd/bsdstg FreeBSD Enterprise Storage Polish BSD User Group What is Enterprise? 2020/02/11 What is Enterprise Storage? FreeBSD Enterprise Storage Polish BSD User Group What is Enterprise? 2020/02/11 What is Enterprise Storage? The wikipedia.org/wiki/enterprise_storage page tells nothing about enterprise. FreeBSD Enterprise Storage Polish BSD User Group What is Enterprise? 2020/02/11 What is Enterprise Storage? The wikipedia.org/wiki/enterprise_storage page tells nothing about enterprise. Actually just re irects to wikipedia.org/wiki/data_storage page. FreeBSD Enterprise Storage Polish BSD User Group What is Enterprise? 2020/02/11 What is Enterprise Storage? The wikipedia.org/wiki/enterprise_storage page tells nothing about enterprise. Actually just re irects to wikipedia.org/wiki/data_storage page. The other wikipedia.org/wiki/computer_data_storage page also oes the same. FreeBSD Enterprise Storage Polish BSD User Group What is Enterprise? 2020/02/11 What is Enterprise Storage? The wikipedia.org/wiki/enterprise_storage page tells nothing about enterprise. Actually just re irects to wikipedia.org/wiki/data_storage page. The other wikipedia.org/wiki/computer_data_storage page also oes the same. The wikipedia.org/wiki/enterprise is just meta page !ith lin"s. FreeBSD Enterprise Storage Polish BSD User Group What is Enterprise? 2020/02/11 Common Charasteristics o Enterprise Storage FreeBSD Enterprise Storage Polish BSD User Group What is Enterprise? 2020/02/11 Common Charasteristics o Enterprise Storage ● Category that inclu es ser$ices/pro ucts esigne &or !arge organizations. FreeBSD Enterprise Storage Polish BSD User Group What is Enterprise? 2020/02/11 Common Charasteristics o Enterprise Storage ● Category that inclu es ser$ices/pro ucts esigne &or !arge organizations. ● Can han le !arge "o!umes o data an !arge num%ers o sim#!tano#s users. FreeBSD Enterprise Storage Polish BSD User Group What is Enterprise? 2020/02/11 Common Charasteristics o Enterprise Storage ● Category that inclu es ser$ices/pro ucts esigne &or !arge organizations. ● Can han le !arge "o!umes o data an !arge num%ers o sim#!tano#s users. ● 'n$olves centra!ized storage repositories such as SA( or NAS e$ices. FreeBSD Enterprise Storage Polish BSD User Group What is Enterprise? 2020/02/11 Common Charasteristics o Enterprise Storage ● Category that inclu es ser$ices/pro ucts esigne &or !arge organizations. ● Can han le !arge "o!umes o data an !arge num%ers o sim#!tano#s users. ● 'n$olves centra!ized storage repositories such as SA( or NAS e$ices. ● )equires more time an experience%expertise to set up an operate. FreeBSD Enterprise Storage Polish BSD User Group What is Enterprise? 2020/02/11 Common Charasteristics o Enterprise Storage ● Category that inclu es ser$ices/pro ucts esigne &or !arge organizations. ● Can han le !arge "o!umes o data an !arge num%ers o sim#!tano#s users. ● 'n$olves centra!ized storage repositories such as SA( or NAS e$ices. ● )equires more time an experience%expertise to set up an operate. ● Generally costs more than consumer or small business storage e$ices. FreeBSD Enterprise Storage Polish BSD User Group What is Enterprise? 2020/02/11 Common Charasteristics o Enterprise Storage ● Category that inclu es ser$ices/pro ucts esigne &or !arge organizations. ● Can han le !arge "o!umes o data an !arge num%ers o sim#!tano#s users. ● 'n$olves centra!ized storage repositories such as SA( or NAS e$ices. ● )equires more time an experience%expertise to set up an operate. ● Generally costs more than consumer or small business storage e$ices. ● Generally o&&ers higher re!ia%i!it'%a"aila%i!it'%sca!a%i!it'. FreeBSD Enterprise Storage Polish BSD User Group What is Enterprise? 2020/02/11 EnterpriCe or EnterpriSe? FreeBSD Enterprise Storage Polish BSD User Group What is Enterprise? 2020/02/11 EnterpriCe or EnterpriSe? DuckDuckGo oes not pro$i e search results count +, Goog!e search &or enterprice !or gi$es ~ 1 )00 000 results. Goog!e search &or enterprise !or gi$es ~ 1 000 000 000 results ,1000 times more). FreeBSD Enterprise Storage Polish BSD User Group What is Enterprise? 2020/02/11 EnterpriCe or EnterpriSe? DuckDuckGo oes not pro$i e search results count +, Goog!e search &or enterprice !or gi$es ~ 1 )00 000 results. Goog!e search &or enterprise !or gi$es ~ 1 000 000 000 results ,1000 times more). ● /ost ictionaries &or enterprice !or sen s you to enterprise term. FreeBSD Enterprise Storage Polish BSD User Group What is Enterprise? 2020/02/11 EnterpriCe or EnterpriSe? DuckDuckGo oes not pro$i e search results count +, Goog!e search &or enterprice !or gi$es ~ 1 )00 000 results. Goog!e search &or enterprise !or gi$es ~ 1 000 000 000 results ,1000 times more). ● /ost ictionaries &or enterprice !or sen s you to enterprise term. ● Given the *+,CE o& many enterprise solutions it coul be enterPRICE 0 FreeBSD Enterprise Storage Polish BSD User Group What is Enterprise? 2020/02/11 EnterpriCe or EnterpriSe? DuckDuckGo oes not pro$i e search results count +, Goog!e search &or enterprice !or gi$es ~ 1 )00 000 results. Goog!e search &or enterprise !or gi$es ~ 1 000 000 000 results ,1000 times more). ● /ost ictionaries &or enterprice !or sen s you to enterprise term. ● Given the *+,CE o& many enterprise solutions it coul be enterPRICE 0 ● 0 or enterpri$e as !ell +. FreeBSD Enterprise Storage Polish BSD User Group What is Enterprise? 2020/02/11 EnterpriCe or EnterpriSe? DuckDuckGo oes not pro$i e search results count +, Goog!e search &or enterprice !or gi$es ~ 1 )00 000 results. Goog!e search &or enterprise !or gi$es ~ 1 000 000 000 results ,1000 times more). ● /ost ictionaries &or enterprice !or sen s you to enterprise term. ● Given the *+,CE o& many enterprise solutions it coul be enterPRICE 0 ● 0 or enterpri$e as !ell +. ● When in oubt just use S $ersion 1 Enterprise. FreeBSD Enterprise Storage Polish BSD User Group 'nternal Solutions 2020/02/11 Internal So!#tions - Filesystems FreeBSD Enterprise Storage Polish BSD User Group 'nternal Solutions 2020/02/11 Internal So!#tions - Filesystems ● UFS 1 classic%mature/traditional small memory &ootprint U('2 3lesytem. FreeBSD Enterprise Storage Polish BSD User Group 'nternal Solutions 2020/02/11 Internal So!#tions - Filesystems ● UFS 1 classic%mature/traditional small memory &ootprint U('2 3lesytem. ○ UFS with So&t Updates (SU) allows Snapshots and dump(8)/restore(8) features. FreeBSD Enterprise Storage Polish BSD User Group 'nternal Solutions 2020/02/11 Internal So!#tions - Filesystems ● UFS 1 classic%mature/traditional small memory &ootprint U('2 3lesytem. ○ UFS with So&t Updates (SU) allows Snapshots and dump(8)/restore(8) features. ○ UFS with Journaled So&t Updates (SU+J) with ultra fast fsck(8) process. FreeBSD Enterprise Storage Polish BSD User Group 'nternal Solutions 2020/02/11 Internal So!#tions - Filesystems ● UFS 1 classic%mature/traditional small memory &ootprint U('2 3lesytem. ○ UFS with So&t Updates (SU) allows Snapshots and dump(8)/restore(8) features. ○ UFS with Journaled So&t Updates (SU+J) with ultra fast fsck(8) process. ● /FS 1 mo ern poole U('2 storage. FreeBSD Enterprise Storage Polish BSD User Group 'nternal Solutions 2020/02/11 Internal So!#tions - Filesystems ● UFS 1 classic%mature/traditional small memory &ootprint U('2 3lesytem. ○ UFS with So&t Updates (SU) allows Snapshots and dump(8)/restore(8) features. ○ UFS with Journaled So&t Updates (SU+J) with ultra fast fsck(8) process. ● /FS 1 mo ern poole U('2 storage. ○ Stable 56S - based on FreeBSD Base System !FS i plementaiton. FreeBSD Enterprise Storage Polish BSD User Group 'nternal Solutions 2020/02/11 Internal So!#tions - Filesystems ● UFS 1 classic%mature/traditional small memory &ootprint U('2 3lesytem. ○ UFS with So&t Updates (SU) allows Snapshots and dump(8)/restore(8) features. ○ UFS with Journaled So&t Updates (SU+J) with ultra fast fsck(8) process. ● /FS 1 mo ern poole U('2 storage. ○ Stable 56S - based on FreeBSD Base System !FS i plementaiton. ○ 7atest 56S - based on !oL/!oF #penZFS repository (use FreeBSD $orts). FreeBSD Enterprise Storage Polish BSD User Group 'nternal Solutions 2020/02/11 Internal So!#tions - Filesystems ● UFS 1 classic%mature/traditional small memory &ootprint U('2 3lesytem. ○ UFS with So&t Updates (SU) allows Snapshots and dump(8)/restore(8) features. ○ UFS with Journaled So&t Updates (SU+J) with ultra fast fsck(8) process. ● /FS 1 mo ern poole U('2 storage. ○ Stable 56S - based on FreeBSD Base System !FS i plementaiton. ○ 7atest 56S - based on !oL/!oF #penZFS repository (use FreeBSD $orts). ○ Differences - http://open-zfs.org/wiki/Feature_/lags - detailed infor ation. FreeBSD Enterprise Storage Polish BSD User Group 'nternal Solutions 2020/02/11 Internal So!#tions - Filesystems ● UFS 1 classic%mature/traditional small memory &ootprint U('2 3lesytem. ○ UFS with So&t Updates (SU) allows Snapshots and dump(8)/restore(8) features. ○ UFS with Journaled So&t Updates (SU+J) with ultra fast fsck(8) process. ● /FS 1 mo ern poole U('2 storage. ○ Stable 56S - based on FreeBSD Base System !FS i plementaiton. ○ 7atest 56S - based on !oL/!oF #penZFS repository (use FreeBSD $orts). ○ Differences - http://open-zfs.org/wiki/Feature_/lags - detailed infor ation. ● F01%E212 1 6reeBSD maintains BSD license 6AT%E2T8 3lesystem implementations. FreeBSD Enterprise Storage Polish BSD User Group 'nternal Solutions 2020/02/11 Internal So!#tions - Frameworks FreeBSD Enterprise Storage Polish BSD User Group 'nternal Solutions 2020/02/11 Internal So!#tions - Frameworks ● GE34 1 6reeBSD's /o ular Disk Transformation 6rame!or".
Recommended publications
  • Gluster Roadmap: Recent Improvements and Upcoming Features
    Gluster roadmap: Recent improvements and upcoming features Niels de Vos GlusterFS co-maintainer [email protected] Agenda ● Introduction into Gluster ● Quick Start ● Current stable releases ● History of feature additions ● Plans for the upcoming 3.8 and 4.0 release ● Detailed description of a few select features FOSDEM, 30 January 2016 2 What is GlusterFS? ● Scalable, general-purpose storage platform ● POSIX-y Distributed File System ● Object storage (swift) ● Distributed block storage (qemu) ● Flexible storage (libgfapi) ● No Metadata Server ● Heterogeneous Commodity Hardware ● Flexible and Agile Scaling ● Capacity – Petabytes and beyond ● Performance – Thousands of Clients FOSDEM, 30 January 2016 3 Terminology ● Brick ● Fundamentally, a filesystem mountpoint ● A unit of storage used as a capacity building block ● Translator ● Logic between the file bits and the Global Namespace ● Layered to provide GlusterFS functionality FOSDEM, 30 January 2016 4 Terminology ● Volume ● Bricks combined and passed through translators ● Ultimately, what's presented to the end user ● Peer / Node ● Server hosting the brick filesystems ● Runs the Gluster daemons and participates in volumes ● Trusted Storage Pool ● A group of peers, like a “Gluster cluster” FOSDEM, 30 January 2016 5 Scale-out and Scale-up FOSDEM, 30 January 2016 6 Distributed Volume ● Files “evenly” spread across bricks ● Similar to file-level RAID 0 ● Server/Disk failure could be catastrophic FOSDEM, 30 January 2016 7 Replicated Volume ● Copies files to multiple bricks ● Similar to file-level
    [Show full text]
  • Storage Virtualization for KVM – Putting the Pieces Together
    Storage Virtualization for KVM – Putting the pieces together Bharata B Rao – [email protected] Deepak C Shettty – [email protected] M Mohan Kumar – [email protected] (IBM Linux Technology Center, Bangalore) Balamurugan Aramugam - [email protected] Shireesh Anjal – [email protected] (RedHat, Bangalore) Aug 2012 LPC2012 Linux is a registered trademark of Linus Torvalds. Agenda ● Problems around storage in virtualization ● GlusterFS as virt-ready file system – QEMU-GlusterFS integration – GlusterFS – Block device translator ● Virtualization management - oVirt and VDSM – VDSM-GlusterFS integration ● Storage integration – libstoragemgmt Problems in storage/FS in KVM virtualization ● Multiple choices for file system and virtualization management ● Lack of virtualization aware file systems ● File systems/storage functionality implemented in other layers of virtualization stack – Snapshots, block streaming, image formats in QEMU ● No well defined interface points in the virtualization stack for storage integration ● No standard interface/APIs available for services like backup and restore ● Need for a single FS/storage solution that works for local, SAN and NAS storage – Mixing different types of storage into a single filesystem namespace GlusterFS ● User space distributed file system that scales to several petabytes ● Aggregates storage resources from multiple nodes and presents a unified file system namespace GlusterFS - features ● Replication ● Striping ● Distribution ● Geo-replication/sync ● Online volume extension
    [Show full text]
  • FOSDEM 2017 Schedule
    FOSDEM 2017 - Saturday 2017-02-04 (1/9) Janson K.1.105 (La H.2215 (Ferrer) H.1301 (Cornil) H.1302 (Depage) H.1308 (Rolin) H.1309 (Van Rijn) H.2111 H.2213 H.2214 H.3227 H.3228 Fontaine)… 09:30 Welcome to FOSDEM 2017 09:45 10:00 Kubernetes on the road to GIFEE 10:15 10:30 Welcome to the Legal Python Winding Itself MySQL & Friends Opening Intro to Graph … Around Datacubes Devroom databases Free/open source Portability of containers software and drones Optimizing MySQL across diverse HPC 10:45 without SQL or touching resources with my.cnf Singularity Welcome! 11:00 Software Heritage The Veripeditus AR Let's talk about The State of OpenJDK MSS - Software for The birth of HPC Cuba Game Framework hardware: The POWER Make your Corporate planning research Applying profilers to of open. CLA easy to use, aircraft missions MySQL Using graph databases please! 11:15 in popular open source CMSs 11:30 Jockeying the Jigsaw The power of duck Instrumenting plugins Optimized and Mixed License FOSS typing and linear for Performance reproducible HPC Projects algrebra Schema Software deployment 11:45 Incremental Graph Queries with 12:00 CloudABI LoRaWAN for exploring Open J9 - The Next Free It's time for datetime Reproducible HPC openCypher the Internet of Things Java VM sysbench 1.0: teaching Software Installation on an old dog new tricks Cray Systems with EasyBuild 12:15 Making License 12:30 Compliance Easy: Step Diagnosing Issues in Webpush notifications Putting Your Jobs Under Twitter Streaming by Open Source Step. Java Apps using for Kinto Introducing gh-ost the Microscope using Graph with Gephi Thermostat and OGRT Byteman.
    [Show full text]
  • Glusterfs Documentation Release 3.8.0
    GlusterFS Documentation Release 3.8.0 Gluster Community Aug 10, 2016 Contents 1 Quick Start Guide 3 1.1 Single Node Cluster...........................................3 1.2 Multi Node Cluster............................................4 2 Overview and Concepts 7 2.1 Volume Types..............................................7 2.2 FUSE................................................... 10 2.3 Translators................................................ 12 2.4 Geo-Replication............................................. 17 2.5 Terminologies.............................................. 19 3 Installation Guide 23 3.1 Getting Started.............................................. 23 3.2 Configuration............................................... 24 3.3 Installing Gluster............................................. 26 3.4 Overview................................................. 27 3.5 Quick Start Guide............................................ 28 3.6 Setup Baremetal............................................. 29 3.7 Deploying in AWS............................................ 30 3.8 Setting up in virtual machines...................................... 31 4 Administrator Guide 33 5 Upgrade Guide 35 6 Contributors Guide 37 6.1 Adding your blog............................................. 37 6.2 Bug Lifecycle.............................................. 37 6.3 Bug Reporting Guidelines........................................ 38 6.4 Bug Triage Guidelines.......................................... 41 7 Changelog 47 8 Presentations 49 i ii GlusterFS
    [Show full text]
  • DMFS - a Data Migration File System for Netbsd
    DMFS - A Data Migration File System for NetBSD William Studenmund Veridian MRJ Technology Solutions NASAAmes Research Center" Abstract It was designed to support the mass storage systems de- ployed here at NAS under the NAStore 2 system. That system supported a total of twenty StorageTek NearLine ! have recently developed DMFS, a Data Migration File tape silos at two locations, each with up to four tape System, for NetBSD[I]. This file system provides ker- drives each. Each silo contained upwards of 5000 tapes, nel support for the data migration system being devel- and had robotic pass-throughs to adjoining silos. oped by my research group at NASA/Ames. The file system utilizes an underlying file store to provide the file The volman system is designed using a client-server backing, and coordinates user and system access to the model, and consists of three main components: the vol- files. It stores its internal metadata in a flat file, which man master, possibly multiple volman servers, and vol- resides on a separate file system. This paper will first man clients. The volman servers connect to each tape describe our data migration system to provide a context silo, mount and unmount tapes at the direction of the for DMFS, then it will describe DMFS. It also will de- volman master, and provide tape services to clients. The scribe the changes to NetBSD needed to make DMFS volman master maintains a database of known tapes and work. Then it will give an overview of the file archival locations, and directs the tape servers to move and mount and restoration procedures, and describe how some typi- tapes to service client requests.
    [Show full text]
  • An Incremental Path Towards a Safer OS Kernel
    An Incremental Path Towards a Safer OS Kernel Jialin Li Samantha Miller Danyang Zhuo University of Washington University of Washington Duke University Ang Chen Jon Howell Thomas Anderson Rice University VMware Research University of Washington LoC Abstract Incremental Progress Tens of Linux Safe Linux Linux has become the de-facto operating system of our age, Millions FreeBSD but its vulnerabilities are a constant threat to service availabil- Hundreds of Singularity ity, user privacy, and data integrity. While one might scrap Thousands Biscuit Linux and start over, the cost of that would be prohibitive due Theseus Thousands RedLeaf seL4 to Linux’s ubiquitous deployment. In this paper, we propose Hyperkernel Safety an alternative, incremental route to a safer Linux through No Type Ownership Functional proper modularization and gradual replacement module by Guarantees Safety Safety Verification module. We lay out the research challenges and potential Figure 1: Our vision and the current state of systems. solutions for this route, and discuss the open questions ahead. security vulnerabilities are reported each year, and lifetime CCS Concepts analysis suggests that the new code added this year has intro- duced tens of thousands of more bugs. ! • Software and its engineering Software verification; While operating systems are far from the only source of se- ! • Computer systems organization Reliability. curity vulnerabilities, it is hard to envision building trustwor- Keywords thy computer systems without addressing operating system correctness. kernel safety, verified systems, reliable systems One attractive option is to scrap Linux and start over. Many ACM Reference Format: past works focus on building more secure and correct op- Jialin Li, Samantha Miller, Danyang Zhuo, Ang Chen, Jon Howell, erating systems from the ground up: ones based on strong and Thomas Anderson.
    [Show full text]
  • Shared File Systems: Determining the Best Choice for Your Distributed SAS® Foundation Applications Margaret Crevar, SAS Institute Inc., Cary, NC
    Paper SAS569-2017 Shared File Systems: Determining the Best Choice for your Distributed SAS® Foundation Applications Margaret Crevar, SAS Institute Inc., Cary, NC ABSTRACT If you are planning on deploying SAS® Grid Manager and SAS® Enterprise BI (or other distributed SAS® Foundation applications) with load balanced servers on multiple operating systems instances, , a shared file system is required. In order to determine the best shared file system choice for a given deployment, it is important to understand how the file system is used, the SAS® I/O workload characteristics performed on it, and the stressors that SAS Foundation applications produce on the file system. For the purposes of this paper, we use the term "shared file system" to mean both a clustered file system and shared file system, even though" shared" can denote a network file system and a distributed file system – not clustered. INTRODUCTION This paper examines the shared file systems that are most commonly used with SAS and reviews their strengths and weaknesses. SAS GRID COMPUTING REQUIREMENTS FOR SHARED FILE SYSTEMS Before we get into the reasons why a shared file system is needed for SAS® Grid Computing, let’s briefly discuss the SAS I/O characteristics. GENERAL SAS I/O CHARACTERISTICS SAS Foundation creates a high volume of predominately large-block, sequential access I/O, generally at block sizes of 64K, 128K, or 256K, and the interactions with data storage are significantly different from typical interactive applications and RDBMSs. Here are some major points to understand (more details about the bullets below can be found in this paper): SAS tends to perform large sequential Reads and Writes.
    [Show full text]
  • Texttext Introduction Into Glusterfs Martin Alfke <Martin.Alfke@Buero20
    Introduction into GlusterFS Martin Alfke <[email protected]> TextText Wednesday, July 7, 2010 1 Agenda • General Information on GlusterFS • Architecture Overview • GlusterFS Translators • GlusterFS Configuration Wednesday, July 7, 2010 2 General Information Wednesday, July 7, 2010 3 General Information I File System Solutions Shared Disk File System •San or Block Access •Mostly used in HA setup (e.g. DRBD) Distributed File System •Network File System •NFS •SMB/CIFS •9P Distributed replicated File System •Replication •HA and offline operation •Coda •MS DFS •MooseFS 4 Wednesday, July 7, 2010 4 General Information II File System Solutions Distributed parallel File System •Setup across multiple servers •HPC Distributed replicated parallel File System •HPC and HA •Cosmos •MogileFS •GPFS (IBM) •GFS (Google) •Hadoop •GlusterFS 5 Wednesday, July 7, 2010 5 Customer Platform Shared Storage Issues Bad performance of NFS kernel stack Limitations of concurrent NFS accesses Customer data already on NFS system Environment and Requirements Debian GNU/Linux Version 4 (etch) – 3 years old Most D f-t p FS need complex data migration Solution has to be expandable 6 Wednesday, July 7, 2010 6 Why GlusterFS ? Decision basics Possibility to run NFS and GlusterFS in parallel No data migration necessary Easy setup Extendable (e.g. new storage nodes) min. Kernel 2.6.3x --> optimization for FUSE context switches ! 7 Wednesday, July 7, 2010 7 Architecture Overview Wednesday, July 7, 2010 8 GlusterFS Basics Hardware Any x86 Hardware Direct attached storage, RAID FC, Infiniband or iSCSI SAN Gigabit or 10 Gigabit network or Infiniband OS Linux, Solaris, OpenSolaris, OS X, FreeBSD ext3 or ext4 Filesystem (tested) Other POSIX compliant filesystems should also work Architecture No Meta-Data Server (fully distributed architecture - Elastic Hash) Replication (RAID 1) Distribution (RAID 0) FUSE (Standard) NFS (unfs3 - depreciated) SMB/CIFS DAV 9 Wednesday, July 7, 2010 9 GlusterFS Architecture Overview I Distribution Replication Images Copyright by Gluster.Inc.
    [Show full text]
  • Red Hat Virtualization 4.3 Planning and Prerequisites Guide
    Red Hat Virtualization 4.3 Planning and Prerequisites Guide Planning for the Installation and Configuration of Red Hat Virtualization 4.3 Last Updated: 2019-05-09 Red Hat Virtualization 4.3 Planning and Prerequisites Guide Planning for the Installation and Configuration of Red Hat Virtualization 4.3 Red Hat Virtualization Documentation Team Red Hat Customer Content Services [email protected] Legal Notice Copyright © 2019 Red Hat, Inc. The text of and illustrations in this document are licensed by Red Hat under a Creative Commons Attribution–Share Alike 3.0 Unported license ("CC-BY-SA"). An explanation of CC-BY-SA is available at http://creativecommons.org/licenses/by-sa/3.0/ . In accordance with CC-BY-SA, if you distribute this document or an adaptation of it, you must provide the URL for the original version. Red Hat, as the licensor of this document, waives the right to enforce, and agrees not to assert, Section 4d of CC-BY-SA to the fullest extent permitted by applicable law. Red Hat, Red Hat Enterprise Linux, the Shadowman logo, JBoss, OpenShift, Fedora, the Infinity logo, and RHCE are trademarks of Red Hat, Inc., registered in the United States and other countries. Linux ® is the registered trademark of Linus Torvalds in the United States and other countries. Java ® is a registered trademark of Oracle and/or its affiliates. XFS ® is a trademark of Silicon Graphics International Corp. or its subsidiaries in the United States and/or other countries. MySQL ® is a registered trademark of MySQL AB in the United States, the European Union and other countries.
    [Show full text]
  • Elo Provides On-Demand Infrastructure with Red Hat
    CUSTOMER CASE STUDY ELO PROVIDES ON-DEMAND INFRASTRUCTURE WITH RED HAT Elo Serviços S.A., a Brazilian payment card company, has grown rapidly since its foundation. However, the market faces increasing competition from innovative financial technology (fintech) startups. To stay competitive, Elo needed an agile, efficient IT environment that would simplify management and speed time to market. With enterprise open source software from Red Hat—such as Red Hat OpenShift Container Platform and Red Hat Ansible Tower—Elo can deploy, manage, and update its customer service and applications faster to stay ahead of tradi- SOFTWARE tional and fintech competition. Red Hat® Enterprise Linux® Red Hat OpenShift® Container Platform Red Hat Ansible® Tower Red Hat Satellite Red Hat Gluster® Storage São Paulo, Brazil FINANCIAL SERVICES SERVICES HEADQUARTERS 115 EMPLOYEES Red Hat Technical Account 115 MILLION CARDS Manager (TAM) IN CIRCULATION Red Hat Consulting “With Red Hat OpenShift, we can plan and complete a proof of concept in 1-2 weeks, HARDWARE BENEFITS then launch what we’ve developed into Dell PowerEdge R730 • Reduced service time to 2S/2U Rack Server production faster to stay ahead of market by speeding server the competition.” deployment from 45 days to 1-2 days ANDERSON AGAPITO I.T. INFRASTRUCTURE MANAGER, ELO • Simplified management with greater automation • Enhanced compliance and security, including PCI-DSS certification • Gained access to expert guidance facebook.com/redhatinc @redhat linkedin.com/company/red-hat redhat.com STAYING COMPETITIVE IN A RAPIDLY GROWING MARKET Elo Serviços S.A., a Brazilian credit card company, was founded in 2011 as a joint venture between three of the country’s leading banks: Banco do Brasil, Bradesco, and Caixa Econômica Federal.
    [Show full text]
  • FS Design Around SMR Approved SNIA Tutorial © 2015 Storage Networking Industry Association
    PRESENTATIONFS Design TITLE Around GOES HERE SMR Seagate’s Journey with EXT4 Adrian Palmer - Seagate SNIA Legal Notice The material contained in this tutorial is copyrighted by the SNIA unless otherwise noted. Member companies and individual members may use this material in presentations and literature under the following conditions: Any slide or slides used must be reproduced in their entirety without modification The SNIA must be acknowledged as the source of any material used in the body of any document containing material from these presentations. This presentation is a project of the SNIA Education Committee. Neither the author nor the presenter is an attorney and nothing in this presentation is intended to be, or should be construed as legal advice or an opinion of counsel. If you need legal advice or a legal opinion please contact your attorney. The information presented herein represents the author's personal opinion and current understanding of the relevant issues involved. The author, the presenter, and the SNIA do not assume any responsibility or liability for damages arising out of any reliance on or use of this information. NO WARRANTIES, EXPRESS OR IMPLIED. USE AT YOUR OWN RISK. FS Design Around SMR Approved SNIA Tutorial © 2015 Storage Networking Industry Association. All Rights Reserved. 2 Abstract SMR Friendly File System SMR is a fundamental drive technology, embraced by drive vendors. SMR changes fundamental assumptions of file system management. This long-held notion of Random-Writes now resembles the write profile of sequential-access tape. Seagate is leading the way in providing a standards compliant IO stack for use with the new drives.
    [Show full text]
  • Advanced File Systems and ZFS
    Advanced File Systems and ZFS Jan Šenolt [email protected] Solaris Engineering May 6, 2021 1 Agenda • Crash Consistency Problem • fsck • Journalling • Log-structured File Systems • Soft-Updates • ZFS 2 Crash Consistency Problem 3 Traditional UNIX File System • appending a new block to the file involves at least 3 writes to different data structures: • block bitmap - allocate the block • inode - update e2di_blocks[], e2di_size • data block - actual payload • what will happen if we fail to make some of these changes persistent? • crash-consistency problem • File System Inconsistency • how to deal with? 4 File System Checker, fsck • a reactive approach • let the inconsistencies happen and try to find (and eventually fix) them later (on reboot) • metadata-only based checks • verify that each allocated block is referenced by exactly one inode • … but what if it is not?? • unable to detect corrupted (missing) user data • does not scale well • O(file system size) • improvements? • check only recently changed data? • … still useful! 5 Journaling, logging 1. start a new transaction 2. write all planned change to the journal 3. make sure that all writes to log completed properly • close the transaction 4. make the actual in-place updates • journal reply • after crash, on reboot • walk the journal, find all complete transactions and apply them 6 Journaling, logging (2) • journal can be a (preallocated) file within the file system or a dedicated device • small circular buffer • UFS: 1MB per 1GB, 64MB max • types of journals • physical - stores
    [Show full text]