Auto-Tiered Storage Conquering I/O Bottlenecks with Hybrid Storage

Total Page:16

File Type:pdf, Size:1020Kb

Auto-Tiered Storage Conquering I/O Bottlenecks with Hybrid Storage Data Sheet Auto-Tiered Storage Conquering I/O Bottlenecks with Hybrid Storage Overview For more than a decade, storage system When data is requested from ZFS, it performance has remained rather stagnant first looks to the ARC; if it is there, it can while drive capacities and application be retrieved extremely fast (typically in performance demands steadily have nanoseconds) and provided back to the increased. The result of this trend is application. The contents of the ARC are an expensive problem: Storage users balanced between the most recently used are forced into buying expensive hard (MRU) and most frequently used (MFU) disk drives (HDDs) to get a moderate data. performance boost (by reducing I/O Level-Two ARC (L2ARC): The L2ARC latency) and/or forced into over-buying lives in SSDs. In concept, it is an extension capacity in order to meet performance of the ARC. Without an L2ARC, data that requirements. could not fit in the ARC would have to be With the advent and decreasing price of retrieved from HDDs when requested. That flash, storage vendors are integrating is when drive speed makes a difference, it into their products to solve this but the performance difference between problem. ZFS technology is leading the “fast” (e.g., 15k-RPM) and “slow” (e.g., industry in its ability to automatically and 7,200-RPM) is in terms of latencies intelligently use flash in a storage system measured as a few milliseconds or that offers the appropriate capacity and several milliseconds; both are dramatically performance capabilities at a total cost slower than ARC accesses measured in that is dramatically lower than most legacy nanoseconds. L2ARC, in SSDs, fits nicely storage systems. between the two—both in terms of price and performance. Buying hundreds of gigabytes of flash is The ZFS Hybrid Storage Pool cheaper than the same capacity of DRAM ZFS is a robust, scalable file system with (though still more expensive today than Hybrid Storage Pools features not available in other file systems HDDs), and flash’s I/O latencies typically available today. One of these revolutionary are measured in microseconds—slower features is the ZFS Hybrid Storage Pool than DRAM but still far faster than even (HSP), which allows you to combine “high-performance” HDDs. The L2ARC is DRAM, SSDs, and spinning HDDs into an populated by data first placed in the ARC accelerated storage medium. Below we will as it becomes apparent that the data might explore each of these offerings. get squeezed out of the ARC, and not every piece of data that existed in ARC will Adaptive Replacement Cache (ARC): make it to the L2ARC (those that do not The ARC lives in DRAM. It is the first would be retrieved from HDDs instead, if destination for all data written to a ZFS requested); the algorithms that manage pool, and it is the fastest (lowest-latency) L2ARC population are automatic and source for data read from a ZFS pool. intelligent. Data Sheet ZFS Intent Log (ZIL): The ZIL is used Hard Disk Drives (HDD): With the to handle synchronous writes—write ARC, L2ARC, and ZIL/SLOG providing operations that are required by protocol the bulk of the performance from a ZFS (e.g., NFS, SMB/CIFS) to be stored in a Hybrid Storage Pool, spinning drives non-volatile location on the storage device are relegated to the job they do well— before they can be acknowledged—to providing lower-performance, higher- ensure data stability. Databases usually density, low-cost storage capacity. Until the require transactions to be on stable day that flash competes with HDDs on a storage devices before continuing, so dollar-per-gigabyte front, the right balance reducing the latency of synchronous writes of DRAM and flash for performance, and has a direct impact on performance. HDDs for capacity, results in a total cost of ownership (TCO) that is less—both initially ZFS can do this by using placing the ZIL and over the long-term—than solving both on a SSD. All writes (whether synchronous or asynchronous) are written into the requirements using all flash or all HDDs. ARC in DRAM, and synchronous writes A New Storage Parameter: are also written to the ZIL before being Working Set Size acknowledged. For legacy storage systems, sizing means Under normal conditions, ZFS regularly determining necessary capacity, IOPS, bundles up all of the recent writes in the and throughput and then performing some ARC and flushes them to the spinning simple math to determine the number of drives—at which point the data in the ZIL is spindles that could provide those numbers. no longer relevant (because it now exists Working Set Size on its long-term, non-volatile destination) As the industry moves towards more and can be replaced by new writes. The sophisticated caching methodologies in ZIL only is read from when synchronous storage systems, a new parameter for writes in the ARC are unable to be written expressing storage needs has become to spinning disk—like after a power failure evident: The Working Set Size (WSS) can be described as the subset of total data or controller failover—at which point ZFS Conclusion reads the ZIL and places that data onto that is actively worked upon (e.g., 500GB the spinning drives as intended. One of this quarter’s sales data out of a total ZFS hybrid storage pools intelligently might compare this concept to non-volatile database of 20TB). combine DRAM, flash, and hard disk drives to achieve the right balance of RAM (NVRAM) from storage vendors like Knowing the WSS makes it possible to cost and performance for any given NetApp, but where NVRAM uses batteries size ARC, L2ARC, and even HDDs more working set, while reducing the need that can wear out and have other issues, accurately, but few applications today have for administrators to constantly monitor write-optimized SLC (single-level cell) flash an awareness of WSS. devices do not need batteries. storage for I/O bottlenecks. And while NVRAM scalability is limited By reducing both read and write latency to available slots, adding SLOGs is as with the use of flash in a ZFS hybrid easy as adding HDDs. Like L2ARC, the storage pool, we end up with a system that ZIL/SLOG is managed automatically and performs far better than legacy storage intelligently by ZFS: Writes that need it are systems, while having a much lower total accelerated, without any additional effort cost of ownership (TCO). by the administrator. Nexenta Systems is the leading supplier of enterprise-class Nexenta Systems, Inc. Nexenta Systems EMEA OpenStorage solutions. Its flagship software-only platform, 455 El Camino Real Camerastraat 8, NexentaStor, delivers high-performance, ultra-scalable, cloud- Santa Clara, CA 95050 1322 BC, Almere and virtualization-optimized storage solutions. www.nexenta.com The Netherlands www.facebook.com/nexenta Copyright © 2012 Nexenta® Systems, twitter.com/nexenta ALL RIGHTS RESERVED Rev. 080712.
Recommended publications
  • Inventors, Technologists and Entrepreneurs
    19 November 2013 We, the undersigned, are a group of inventors, technologists and entrepreneurs. Many of us have founded technology businesses; we have invented many of the protocols, systems and devices that make the Internet work, and we are collectively listed as the inventors on over 150 patents. We write to you today about the U.S. patent system. That system is broken. Based on our experiences building and deploying new digital technologies, we believe that software patents are doing more harm than good. Perhaps it is time to reexamine the idea, dating from the 1980s, that government-issued monopolies on algorithms, protocols and data structures are the best way to promote the advancement of computer science. But that will be a complex task, and one we don't expect to happen quickly. Unfortunately, aspects of the problem have become so acute they must be addressed immediately. Broad, vague patents covering software-type inventions—some of which we ourselves are listed as inventors on—are a malfunctioning component of America's inventive machinery. This is particularly the case when those patents end up in the hands of non-practicing patent trolls. These non-practicing entities do not make or sell anything. Their exploitation of patents as a tool for extortion is undermining America’s technological progress; patent trolls are collecting taxes on innovation by extracting billions of dollars in dubious licensing fees, and wasting the time and management resources of creative businesses. Many of us would have achieved much less in our careers if the trolling problem had been as dire in past decades as it is now.
    [Show full text]
  • Education and Research Solutions Brief
    Education and Research Solutions Brief Data Integrity, Performance, Flexibility, and Lower Total Cost of Storage Whether a university department or a large research lab, higher education and research institutions need to provide an IT infrastructure that will allow users to store, retrieve, and analyze data generated by their user groups whether on campus, around the world, or from outer space. Enter: Nexenta Nexenta software-only products solve accessibility, storage performance, and data management In summary: bottlenecks and do so cost effectively. Nexenta provides a cost optimized, hardware independent Why Choose Nexenta? storage platform to safely keep data and make it accessible from anywhere. It delivers superior performance running on industry standard x86 reference architectures from leading hardware • Total Flexibility on hardware choice vendors such as Dell, HP, SuperMicro, Cisco, Quanta, and others. • Lowest total cost for storage • Highest levels of data integrity • Improved data performance With Nexenta, Research Institutes can meet higher levels of: Data Integrity: the large, complex, and sometimes sensitive data sets at higher education and research institutions require the utmost in data reliability. Using the proven ZFS file system, NexentaStor provides unparalleled levels of reliability against data corruption and bit rot Revolutionize your data storage. which has become a growing issue as the amount of data has exploded; and ZFS copy-on- write delivers 100% data integrity. Choose Nexenta Performance: User groups at these institutions put high demands on their data. High levels of Software-Defined Storage memory caching in its ZFS technology allow NexentaStor to deliver superior I/O for the fastest Request a proof-of-concept at access to content.
    [Show full text]
  • Nexentastor 5.X Reference Architecture- Supermicro
    NexentaStor 5.x ReFerence Architecture May 2019 Copyright © 2019 Nexenta Systems, ALL RIGHTS RESERVED www.nexenta.com NexentaStor 5.x ReFerence Architecture- Supermicro TabLe oF Contents Table of Contents ............................................................................................................................ 2 Preface ............................................................................................................................................ 3 Intended Audience ................................................................................................................................................................................. 3 Comments .................................................................................................................................................................................................. 3 Copyright, Trademarks, and Compliance ..................................................................................................................................... 3 1 Supermicro X11 Reference Architectures ................................................................................... 4 1.1 Supermicro X11 All-Flash Configurations ....................................................................................................................... 4 1.1.1 Supermicro X11 All-Flash – 24 Bay SC216 .................................................................................................................... 4 1.2 Supermicro X11 Hybrid Configurations
    [Show full text]
  • Nexentastor Installation Guide 4.0.4 FP1
    NexentaStor Installation Guide 4.0.4 FP1 Date: June, 2016 Subject: NexentaStor Installation Guide Software: NexentaStor Software Version: 4.0.4 FP1 Part Number: 2000-nxs-4.0.4 FP1-0000012-B Copyright © 2016 Nexenta Systems, ALL RIGHTS RESERVED www.nexenta.com NexentaStor Installation Guide Copyright © 2016 Nexenta SystemsTM, ALL RIGHTS RESERVED Notice: No part of this publication may be reproduced or transmitted in any form or by any means, electronic or mechanical, including photocopying and recording, or stored in a database or retrieval system for any purpose, without the express written permission of Nexenta Systems (hereinafter referred to as “Nexenta”). Nexenta reserves the right to make changes to this document at any time without notice and assumes no responsibility for its use. Nexenta products and services only can be ordered under the terms and conditions of Nexenta Systems’ applicable agreements. All of the features described in this document may not be available currently. Refer to the latest product announcement or contact your local Nexenta Systems sales office for information on feature and product availability. This document includes the latest information available at the time of publication. Nexenta, NexentaStor, NexentaEdge, and NexentaConnect are registered trademarks of Nexenta Systems in the United States and other countries. All other trademarks, service marks, and company names in this document are properties of their respective owners. Product Versions Applicable to this Documentation: Product Versions supported NexentaStorTM 4.0.4 FP1 Copyright © 2016 Nexenta Systems, ALL RIGHTS RESERVED ii www.nexenta.com NexentaStor Installation Guide Contents Preface . .vii 1 Introduction . .1 About NexentaStor . .1 NexentaStor Advantages .
    [Show full text]
  • Nexentastor Hardware Reference Architectures
    NexentaStor 5.x Hardware Certification List (HCL) Nexenta Certification Team Dec 2019 Table of Contents 1 Overview .................................................................................................................................. 7 1.1 Introduction ............................................................................................................................................................................... 7 1.2 NexentaStor Solutions ........................................................................................................................................................... 7 1.2.1 Reference Architectures (RA) ............................................................................................................................................. 7 1.2.2 Certified Solutions (CS) .......................................................................................................................................................... 7 1.3 Common Pre-Requisites ....................................................................................................................................................... 8 2 Cisco Reference Architectures ................................................................................................... 9 2.1 Cisco All-Flash Configurations ........................................................................................................................................... 9 2.1.1 Cisco C240 and HGST 2U24 All-Flash .............................................................................................................................
    [Show full text]
  • Meet Nexenta
    Who we are Meet Nexenta is the Global Leader in Open Source-driven Software-Defined Storage. We lead the market in combining our Open Source-driven Software with the most innovative industry standard hardware to provide you the ability to maximize the potential of the storage in your data center. Nexenta. Today, Nexenta Software-Defined Storage manages over 2.0 Exabytes of storage. What we do “Our decision to go with Nexenta Nexenta provides you with the functionality, flexibility and performance that you want at a significantly lower TCO than hardware appliance vendors, will save the university around $9 delivering greater value that satisfies both your current and future data million over a five year period. ‘Tier 1’ center needs. By providing a software-only storage solution, including a rich storage vendors would have charged feature set across your complete protocol’s needs (block, file, and object), Nexenta enables you to deliver a Software-Defined infrastructure to your 3 to 3.5 times more than the amount legacy and next generation applications, while giving you the freedom to negotiated with Nexenta and yet the choose the platform you want to run on. functionality was basically the same.” What makes us different Patrick J.A. Hopewell Our vision. It’s focused on promoting and delivering the true benefits Director, Enterprise Infrastructure Solutions of Software-Defined Everything. We use our vision to promote and University of Toronto drive open partnerships with our customers and partners, including server hardware, networking, SSDs, HDDs, cloud platforms, application infrastructure, business analytics, and everything else considered part of an IT infrastructure.
    [Show full text]
  • Introducing a New Product
    illumos SVOSUG Update Presented by Garrett D'Amore Nexenta Systems, Inc. August 26, 2010 What's In A Name? illumos = illum + OS = “Light + OS” Light as in coming from the Sun... OS as in Operating System Note: illumos not Illumos or IllumOS “illumos” trademark application in review. Visual branding still under consideration. Not All of OpenSolaris is Open Source ● Critical components closed source – libc_i18n (needed for working C library) – NFS lock manager – Portions of crypto framework – Numerous critical drivers (e.g. mpt) ● Presents challenges to downstream dependents – Nexenta, Belenix, SchilliX, etc. – See “Darwin” and “MacOS X” for the worst case What's Good ● The Technology! – ZFS, DTrace, Crossbow, Zones, etc. ● The People – World class engineers! – Great community of enthusiasts – Vibrant ecosystem ● The Code is Open – Well most of it, at least illumos – the Project ● Derivative (child) of OS/Net (aka ON) – Solaris/OpenSolaris kernel and foundation – 100% ABI compatible with Solaris ON – Now a real fork of ON, but will merge when code available from Oracle ● No closed code – Open source libc, kernel, and drivers! ● Repository for other “experimental” innovations – Can accept changes from contributors that might not be acceptable to upstream illumos – the Ecosystem ● illumos-gate is just ON – Focused on “Core Foundation Blocks” – Flagship project ● Expanding to host other affiliated projects – Umbrella organization – X11 components? – Desktop components? – C++ Runtime? – Distributions? illumos – the Community ● Stands independently
    [Show full text]
  • Three Open Source Network Storage Software
    Three Open Source Network Storage Software Whether you’re looking to create a scalable storage solution for your organization or a simple and easy to setup storage box for your home or small office, here are three packages that will meet all your network storage requirements. You can either download their latest editions or take them from this month’s PCQ DVD. - Sandeep Koul NAS4Free This software is actually related to FreeNAS (the one we’ve done a detailed implementation guide on in this issue), which is probably why it has a similar sounding name. In fact, NAS4Free is a direct continuation of efforts that went into developing the original FreeNAS code, which started in 2005 and continued till 2011 by the name FreeNAS. Post 2011, these two packages were developed separately by two different communities. Like FreeNAS, NAS4Free also comes loaded with features like sharing across Windows, Apple, and UNIX-like systems, includes ZFS, software RAID (0,1,5), disk encryption, S.M.A.R.T / email reports etc. It supports all common network storage protocols including CIFS (samba), FTP, NFS, TFTP, AFP, RSYNC, Unison, iSCSI, UPnP, Bittorent (initiator and target). All these features can be configured through the software’s Web interface. The package is an embedded Open Source Network-Attached Storage distribution and is based on FreeBSD. Download from: www.nas4free.org/ OpenMediaVault The great feature about this storage package is the modular design of its framework, due to which its functionality can be enhanced via plug-ins. OpenMediaVault is based pcquest.com twitter.com/pcquest facebook.com/pcquest linkd.in/pcquest [email protected] DECEMBER 2012 PCQUEST 51 COVER STORY on Debian Linux and contains services like SSH, (S)FTP, with the pricing determined on a per-terabyte and per- SMB/CIFS, DAAP media server, RSync, BitTorrent client server licensing basis.
    [Show full text]
  • A Technical Summary
    A Technical Summary Table A.1. Kerberos administrative ACL (MIT) Allow Deny Operation a A Add principals or policies d D Delete principals or policies m M Modify principals or policies c C Change passwords i I Inquiry the database l L List principals or policies s S Set key for principals * or x Allow all privileges Table A.2. Kerberos kadmin password policies (MIT) Option Meaning -history Minimum count of unusable old passwords -maxlife Maximum allowed lifetime -minclasses Minimum character classes -minlength Minimum password length -minlife Minimum lifetime 380 A Technical Summary Table A.3. LDAP access control list levels Level Privilege Coding Explanation none =0 No access at all auth =x Permits authentication attempt compare =cx Permits comparison search =scx Permits search filter application read =rscx Permits search result inspection write =wrscx Permits modification or deletion Table A.4. The OpenAFS access control list attributes Permission Meaning l List contents (lookup) i Create new files or directories (insert) d Delete files or directories a Change ACL attributes of directories (administer) r Read contents of files w Modify contents of files (write) k Lock files for reading read Equivalent to rl write Equivalent to rlidwk (no administrative rights) all Equivalent to rlidwka none Remove all ACL permissions Table A.5. The OpenAFS @sys names (excerpt) @sys Architecture alpha dux40 Digital UNIX 4 on an Alpha alpha dux50 Digital UNIX 5 on an Alpha i386 linux24 Linux Kernel 2.4 on Intel and compatible i386 linux26 Linux Kernel 2.6 on Intel and compatible i386 nt40 Microsoft Windows NT and later on Intel and compatible ppc darwin70 Apple MacOS X 10.3 on a PowerPC Macintosh ppc darwin80 Apple MacOS X 10.4 on a PowerPC Macintosh ppc darwin90 Apple MacOS X 10.5 on a PowerPC Macintosh rs aix52 IBM AIX 5.2 on a pSeries rs aix53 IBM AIX 5.3 on a pSeries sgi 65 SGI Irix 6.5 on a MPIS x86 darwin80 Apple MacOS X 10.4 on an Intel Macintosh x86 darwin90 Apple MacOS X 10.5 on an Intel Macintosh A Technical Summary 381 Table A.6.
    [Show full text]
  • Nexentaedge Datasheet
    NexentaEdge Datasheet Multi-Service Scale-Out Storage Software NexentaEdge provides scale-out Block, File and Object (S3 & Swift) storage services delivering cluster-wide inline deduplication and compression at chunk level to deliver enterprise grade data integrity. Quick Erasure Coding and shared namespace between NFS and object data provide best in class performance with low touch operating model. Next Generation Scale-Out Architecture NexentaEdge is a truly distributed, scale-out architecture, Standard servers consisting of four or more physicalservers interconnected using NexentaEdge NexentaEdge NexentaEdge a dedicated 10 Gigabit Ethernet (10 GbE) network for cluster communication. The connected servers form a cluster that maintains NexentaEdge NexentaEdge NexentaEdge redundancy and resilience of data throughout the system using strong All-flash, hybrid or all-disk cryptographic checksums for data integrity, and replication NexentaEdge NexentaEdge NexentaEdge technology to ensure hardware-level redundancy. NexentaEdge runs on shared nothing clusters of industry standard Linux servers. See the NexentaEdge Configuration Guidelines for Multi-service Scale-Out Block, File & Object more information. Supporting Platforms for Converged Infrastructures NexentaEdge is ideally suited to support use cases from Enterprise and Next Generation Clouds, Object Applications, and large Object-Based Archives. NexentaEdge clusters can be deployed as either all-flash, hybrid, or all-disk configurations. Enterprise Clouds Next Gen Clouds & Object Apps Object Archive All-Flash Hybrid All HDD iSCSI, Native Block, NFS, S3 and Swift S3, Swift, NFS High Performance Scalability Inline Deduplication & Compression Quick EC Ease of operation Data integrity Visit us online at nexenta.com NexentaEdge System Requirements & Feature Highlights Key Features of NexentaEdge NexentaEdge software provides flexibility in hardware configuration and brings unparalleled simplicity to your daily user experience regardless of your deployment size.
    [Show full text]
  • The Cernvm Project a New Approach to Software Distribution
    The CernVM Project A new approach to software distribution Carlos Aguado [email protected] Jakob [email protected] Predrag [email protected] Outline • Motivation • Mission • Requirements • Service architecture • Conclusions CernVM – A virtual machine for LHC experiments HEPiX Umeå May 27, 2009 Outline • Motivation • Mission • Requirements • Service architecture • Conclusions CernVM – A virtual machine for LHC experiments HEPiX Umeå May 27, 2009 Motivation • LHC experiment frameworks are a complex world – Heavy, multiplatform (arch+OS+compiler) – Grid and development environments • Industry trend to Multi/Many‐Core CPU architectures • Demand for scalable, parallel applications (simulation, reconstruction, analysis) – Large architectural impact, different programming model CernVM – A virtual machine for LHC experiments HEPiX Umeå May 27, 2009 Outline • Motivation • Mission • Requirements • Service architecture • Conclusions CernVM – A virtual machine for LHC experiments HEPiX Umeå May 27, 2009 Project mission • Portable Analysis Environment using Virtualization Technology (WP9) – Approved in 2007 (2+2 years) as R&D activity in CERN/PH Department • Project goals: – Provide a complete, portable and easy to configure user environment for developing and running LHC data analysis – Reduce effort to install, maintain and keep up to date the experiment software – Decouple application lifecycle from evolution of system infrastructure – Lower the cost of software development by reducing the number of platforms CernVM – A virtual machine for LHC experiments
    [Show full text]
  • Nexenta, Openstorage and Commercial Open Source
    Nexenta, OpenStorage and Commercial Open Source Anil Gulecha Developer / Community Lead, Nexenta Core Platform What we'll cover Nexenta Core Platform OpenSolaris Ubuntu Nexenta's difference Distribution Creation Derivatives Nexenta.org Commercial Open Source The OpenStorage Industry NexentaStor www.nexentastor.org Nexenta Core Platform Nexenta is a combination of two great technologies: OpenSolaris : The enterprise class FOSS kernel Ubuntu userland : The most popular *nix userland and packaging framework Started in 2005 as a desktop distribution, based on Ubuntu. Proved hard to support due to small dev team Starting from early 2008, the project only concentrated on a command-line only core platform, which became known as the Nexenta Core Platform. The focus was to create the one of the best server platform out there. NCP: The OpenSolaris side Nexenta is based on the opensolaris kernel. It thus has goodies like The ZFS filesystem, which has revolutionized the open source file systems arena. Dtrace, which provides administrators and developers with the ability to analyze the system in a fine-grained manner. Zones, the lightweight virtualization which allows for opensolaris (and linux) virtualization on Nexenta. Nexenta maintains patches of it's own improvements to the OpenSolaris kernel, which includes patches providing iSCSI on ZFS volumes, and CIFS performance. To fit into the Ubuntu/Debian model, the opensolaris kernel is packaged as .deb packages, for ex: sunwckr : the core kernel runtime package sunwzfs : the zfs package NCP: The Ubuntu side All userland packages, which include shell commands, server packages, tools and utilities are ports of upstream Ubuntu packages. Ubuntu has a lot of packages.
    [Show full text]