Implementing Iscsi Boot with Oracle Solaris on an Oracle ZFS Storage Appliance

Total Page:16

File Type:pdf, Size:1020Kb

Implementing Iscsi Boot with Oracle Solaris on an Oracle ZFS Storage Appliance An Oracle Technical White Paper January 2014 Implementing iSCSI Boot with Oracle Solaris on the Oracle ZFS Storage Appliance Implementing iSCSI Boot with Oracle Solaris on an Oracle ZFS Storage Appliance Table of Contents Introduction ................................................................................................ 4 Benefits of Booting Using iSCSI ............................................................. 4 Differences from a Network PXE Boot .................................................... 4 iSCSI Boot Configuration Requirements ..................................................... 5 Oracle Solaris SPARC Platforms ............................................................ 5 x86 Platforms ......................................................................................... 5 Oracle Solaris Operating Systems .......................................................... 5 Network Configuration and iSCSI Protocol Parameters .......................... 6 Controlling the Boot Process on a SPARC Server .................................. 6 Controlling the Boot Process on an x86-based Server ............................ 6 DHCP Configuration Considerations ....................................................... 7 Required Configuration Information for Setting Up an iSCSI Connection ............................................................................................. 7 Oracle Solaris installation ................................................................................. 8 Oracle Solaris boot ........................................................................................... 8 Configuring the Oracle ZFS Storage Appliance for iSCSI Boot ................... 8 Creating a Project ................................................................................... 9 Creating iSCSI Target and iSCSI Target Groups .................................. 10 Creating iSCSI Initiators and iSCSI Initiator Groups ............................. 11 Creating iSCSI Boot LUNs .................................................................... 12 Installing Oracle Solaris on the Server ...................................................... 14 Installing Oracle Solaris on a SPARC Platform ..................................... 14 Installing Oracle Solaris on an x86 Platform ......................................... 15 Installing the iSCSI Component (Platform-independent) ....................... 15 Configuring the Server for iSCSI Boot ...................................................... 19 Rebooting on the Oracle Solaris SPARC Platform ................................ 19 Rebooting on the x86 Platform ............................................................. 20 Oracle Solaris Setup ............................................................................. 22 Best Practices for Deploying iSCSI Boot .................................................. 23 Network Topology ................................................................................. 23 Using Multipathing ................................................................................ 23 Using DHCP ......................................................................................... 24 2 Implementing iSCSI Boot with Oracle Solaris on an Oracle ZFS Storage Appliance Troubleshooting iSCSI Configuration Problems ........................................ 24 No Network Connection Configured During Oracle Solaris Installation ............................................................................................ 24 Need to Verify iSCSI Targets' Access from Initiator .............................. 24 Too Many LUNs Showing Up................................................................ 26 Appendix A: Manually Configuring a Network Connection ........................ 27 Appendix B: DHCP Setup Tips ................................................................. 29 Setting up for configuration with a fixed server IP address ............................ 29 Setting up DHCP to provide iSCSI boot disk info and initiator IQN ............... 29 Using DHCP for iSCSI Boot in Coexistence with Automated Installer ........... 30 Appendix C: Configuration Information Handling ...................................... 32 Appendix D: References ........................................................................... 33 Table of Figures Figure 1. Required information for iSCSI setup ........................................... 8 Figure 2. Creating a new project on the Oracle ZFS Storage Appliance......................................................................................... 9 Figure 3. Setting project properties ............................................................. 9 Figure 4. Project LUN defaults .................................................................. 10 Figure 5. Creating the iSCSI target ........................................................... 11 Figure 6. iSCSI targets and target groups ................................................. 11 Figure 7. Registering an iSCSI Initiator IQN ............................................. 12 Figure 8. ISCSI Initiator IQN Groups setup ............................................... 12 Figure 9. Creating an iSCSI boot LUN ...................................................... 13 Figure 10. Created boot LUNs displayed in the Oracle ZFS Storage Appliance BUI .................................................................. 13 Figure 11. iSCSI LUN option .................................................................... 15 Figure 12. Specifying iSCSI configuration information to locate the target LUN ..................................................................................... 16 Figure 13. Available disks for Oracle Solaris installation ........................... 17 Figure 14. LUN GUID information displayed in the Oracle ZFS Storage Appliance BUI .................................................................. 17 Figure 15. Selected boot disk details ........................................................ 18 Figure 16. Analytics to show iSCSI install I/O ........................................... 18 Figure 17. Starting NIC iSCSI setup ......................................................... 20 Figure 18. NIC setup; selecting the primary boot port ............................... 21 Figure 19. NIC setup; entering iSCSI configuration information ................ 22 Figure 20. NIC setup; saving changes ...................................................... 22 Figure 21. Oracle Solaris install on iSCSI LUN flowchart .......................... 32 3 Implementing iSCSI Boot with Oracle Solaris on an Oracle ZFS Storage Appliance Introduction The iSCSI boot feature of Oracle Solaris enables you to load and start Oracle Solaris over the network from a remote location. The iSCSI boot feature supports booting from both SPARC-based and x86-based hardware systems. iSCSI (Internet Small Computer System Interface) boot is typically initiated by the server while the boot image resides on an iSCSI target that is attached to the network. Because the feature uses a standard Ethernet-based infrastructure, only standard network infrastructure components are required. This paper describes how to set up the Oracle ZFS Storage Appliance to use iSCSI LUNs from which to boot Oracle Solaris, how to install Oracle Solaris on the configured iSCSI LUN, and how to configure the server boot environment to boot from the installed Oracle Solaris image. This paper also details the use of Dynamic Host Configuration Protocol (DHCP) services. Using DHCP simplifies the Oracle Solaris installation process and it can be configured to supply the iSCSI boot information needed by the server’s hardware boot process. Note: In this paper a Network Interface Adapter (NIC) of the Intel 1000/Pro server series is used for the x86-hardware-based examples. Benefits of Booting Using iSCSI An iSCSI boot solution provides significant benefits: Reduces data center footprint and facility costs – Booting using iSCSI enables you to use diskless servers and blade servers, which take up less space, consume less power, and require less cooling. Lowers administrative overhead – All operating system storage is centrally provisioned and managed from the Oracle ZFS Storage Appliance. A server can be easily replaced by remapping its corresponding boot LUN to a new server. If the new server has the same hardware profile as the server being replaced, the server will boot the operating system from the iSCSI LUN without requiring reconfiguration. Snapshots and clones of operating system images can be created and mapped to new servers on the SAN with just a few clicks of a mouse, simplifying migration and scalability tasks. Facilitates disaster and server failure recovery – By installing operating systems on the Oracle ZFS Storage Appliance rather than individual servers, you can take advantage of the data protection and redundancy features of the Oracle ZFS Storage Appliance to help reduce downtime normally caused by maintenance and fault outages. Operating system images can be protected using snapshots and clones, backed up using Network Data Management Protocol (NDMP) or images replicated to a second ZFS Storage Appliance on an off-site location. Differences from a Network PXE Boot Using iSCSI boot on a SPARC-based system to boot over the network differs from a typical SPARC network PXE (Pre-boot eXecution Environment) boot by its streamlined process. The network boot process is a multistep process; it requires loading a boot environment
Recommended publications
  • Working with DHCP in Oracle® Solaris 11.2
    ® Working With DHCP in Oracle Solaris 11.2 Part No: E36812 July 2014 Copyright © 1999, 2014, Oracle and/or its affiliates. All rights reserved. This software and related documentation are provided under a license agreement containing restrictions on use and disclosure and are protected by intellectual property laws. Except as expressly permitted in your license agreement or allowed by law, you may not use, copy, reproduce, translate, broadcast, modify, license, transmit, distribute, exhibit, perform, publish, or display any part, in any form, or by any means. Reverse engineering, disassembly, or decompilation of this software, unless required by law for interoperability, is prohibited. The information contained herein is subject to change without notice and is not warranted to be error-free. If you find any errors, please report them to us in writing. If this is software or related documentation that is delivered to the U.S. Government or anyone licensing it on behalf of the U.S. Government, the following notice is applicable: U.S. GOVERNMENT END USERS. Oracle programs, including any operating system, integrated software, any programs installed on the hardware, and/or documentation, delivered to U.S. Government end users are "commercial computer software" pursuant to the applicable Federal Acquisition Regulation and agency-specific supplemental regulations. As such, use, duplication, disclosure, modification, and adaptation of the programs, including any operating system, integrated software, any programs installed on the hardware, and/or documentation, shall be subject to license terms and license restrictions applicable to the programs. No other rights are granted to the U.S. Government. This software or hardware is developed for general use in a variety of information management applications.
    [Show full text]
  • EMC Host Connectivity Guide for Oracle Solaris
    Dell EMC Host Connectivity Guide for Oracle Solaris P/N 300-000-607 REV 56 MAY 2020 Copyright © 2007 – 2020 Dell Inc. or its subsidiaries. All rights reserved. Dell believes the information in this publication is accurate as of its publication date. The information is subject to change without notice. THE INFORMATION IN THIS PUBLICATION IS PROVIDED “AS-IS.” DELL MAKES NO REPRESENTATIONS OR WARRANTIES OF ANY KIND WITH RESPECT TO THE INFORMATION IN THIS PUBLICATION, AND SPECIFICALLY DISCLAIMS IMPLIED WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. USE, COPYING, AND DISTRIBUTION OF ANY DELL SOFTWARE DESCRIBED IN THIS PUBLICATION REQUIRES AN APPLICABLE SOFTWARE LICENSE. Dell Technologies, Dell, EMC, Dell EMC and other trademarks are trademarks of Dell Inc. or its subsidiaries. Other trademarks may be the propertyof their respective owners. Published in the USA. Dell EMC Hopkinton, Massachusetts 01748-9103 1-508-435-1000 In North America 1-866-464-7381 www.DellEMC.com 2 Dell EMC Host Connectivity Guide for Oracle Solaris CONTENTS Preface ....................................................................................................................................... 13 Part 1 Connecting Solaris to Dell EMC Storage Chapter 1 Solaris Operating System Solaris operating system overview........................................................................ 20 Multipathing software ........................................................................................... 21 MPxIO/STMS ...............................................................................................
    [Show full text]
  • Fibre Channel and Iscsi Configuration Guide for the Data ONTAP® 8.0 Release Family
    Fibre Channel and iSCSI Configuration Guide for the Data ONTAP® 8.0 Release Family NetApp, Inc. 495 East Java Drive Sunnyvale, CA 94089 U.S. Telephone: +1 (408) 822-6000 Fax: +1 (408) 822-4501 Support telephone: +1 (888) 463-8277 Web: www.netapp.com Feedback: [email protected] Part number: 215-08164_A0 August 2013 Table of Contents | 3 Contents iSCSI configurations .................................................................................... 6 Single-network HA pair in an iSCSI SAN .................................................................. 6 Multi-network HA pair in an iSCSI SAN ................................................................... 7 Direct-attached single-controller configurations in an iSCSI SAN ............................ 8 VLANs for iSCSI configurations ................................................................................ 9 Static VLANs ................................................................................................ 10 Dynamic VLANs ........................................................................................... 10 Fibre Channel configurations .................................................................... 11 FC onboard and expansion port combinations .......................................................... 11 Fibre Channel supported hop count .......................................................................... 12 Fibre Channel supported speeds ................................................................................ 13 Fibre Channel switch
    [Show full text]
  • Pf3e Index.Pdf
    INDEX Note: Pages numbers followed by f, n, priority-based queues, 136–145 or t indicate figures, notes, and tables, match rule for queue assignment, respectively. 137–138 overview, 134–135 Symbols performance improvement, 136–137 # (hash mark), 13, 15 queuing for servers in DMZ, ! (logical NOT) operator, 42 142–144 setting up, 135–136 A on FreeBSD, 135–136 on NetBSD, 136 Acar, Can Erkin, 173 on OpenBSD, 135 ACK (acknowledgment) packets transitioning to priority and class-based bandwidth allocation, queuing system, 131–133 139–140 anchors, 35–36 HFSC algorithm, 124, 126, 142 authpf program, 61, 63 priority queues, 132, 137–138 listing current contents of, 92 two-priority configuration, loading rules into, 92 120–121, 120n1 manipulating contents, 92 adaptive.end value, 188 relayd daemon, 74 adaptive firewalls, 97–99 restructuring rule set with, 91–94 adaptive.start value, 188 tagging to help policy routing, 93 advbase parameter, 153–154 ancontrol command, 46n1 advskew parameter, 153–154, 158–159 antispoof tool, 27, 193–195, 194f aggressive value, 192 ARP balancing, 151, 157–158 ALTQ (alternate queuing) framework, atomic rule set load, 21 9, 133–145, 133n2 authpf program, 59–63, 60 basic concepts, 134 basic authenticating gateways, class-based bandwidth allocation, 60–62 139–140 public networks, 62–63 overview, 135 queue definition, 139–140 tying queues into rule set, 140 B handling unwanted traffic, 144–145 bandwidth operating system-based queue actual available, 142–143 assignments, 145 class-based allocation of, 139–140 overloading to
    [Show full text]
  • Iscsi Testing: What Are the Test Challenges Under the Hood of a 10 Gb Iscsi Storage Product Certification?
    iSCSI testing: What are the test Challenges Under the Hood of a 10 Gb iSCSI Storage Product Certification? Dr. M. K. Jibbe Distinguished Engineer Manager and Technical Lead of Test Architect and Product Certification in India LSI Corporation (Engenio Storage Group) Storage Developer Conference 2009 © 2009 Insert Copyright Information Here. All Rights Reserved. 1 Abstract iSCSI RAID Storage Testing The certification of a 10 Gb iSCSI RAID Storage System elicits a lot of challenges at the development level and the Test / Quality Assurance level. The challenges are due to the fact that a 10 Gb iSCSI is a newly deployed iSCSI host interface in the RAID Storage environment. As a result the size of a development module level test should be designed very carefully to establish a test coverage beyond basic implementation verification, standard RAID testing, or the iSCSI plug fest. These module level tests must tackle the test time windows associated with the following iSCSI characteristics: NIC vs. CNA Device discovery, 10 GB switch traffic control and congestion, Security mechanisms with different Operating systems, Operational parameters associated with I/O retries and recovery Management, Administration, and Integration with Storage products Design For Testability “DFT” mechanisms Diagnostics, problem Isolations IPV4 vs. IPV6 However a number of the module tests above can be leveraged from the certification a 1 Gb iSCSI RAID products. There are specific features such as backup, snapshot, remote mirroring, and cluster application compatibility that must be supported by the RAID product and must be verified during the testing of the RAID controller host interface. Storage Developer Conference 2009 © 2009 Insert Copyright Information Here.
    [Show full text]
  • Install and Setup DHCP Service
    Install and setup DHCP Service Synopsis: DHCP or Dynamic Host Configuration Protocol allows for automatic management of IP addresses on your internal network. By providing DHCP on your LAN, you eliminate the bother of having to configure manually PCs, laptops, or smart phones. They will automatically get an IP address and be configured with default gateway and DNS server information. It is important that only one DHCP server be active on a network. So, check carefully any other network devices on your network, especially wireless Access Points, to be sure their DHCP service is disabled. I. Installation Log into the Freebsd system as root. (Assumes networking is already configured.) # # cd /usr/ports/net/isc-dhcp44-server && make install clean # II. Startup Configuration Enable DHCP daemon at startup and assign the internal LAN (sk0 in this project) interface; # # echo 'dhcpd_enable="YES"' >> /etc/rc.conf # echo 'dhcpd_ifaces="sk0"' >> /etc/rc.conf # III. Setup configuration file # # cd /usr/local/etc # If there is a dhcpd.conf file, rename it; # # mv dhcpd.conf dhcpd.conf.original # Copy the following text, modify for your personal use, and paste it into a new dhcpd.conf file. You will need to modify the subnet (i.e., your network zero) and mask, the IP pool range, the DNS server (in this case 8.8.8.8 is the Google DNS server), the domain name, the router/gateway and the broadcast address (i.e., your network 255). The settings shown assumes our LAN subnet will be the private range 192.168.1.0 to 192.168.1.255. This machine will have LAN address configured as 192.168.1.254 and will be the default gateway for the LAN network.
    [Show full text]
  • The Complete Freebsd
    The Complete FreeBSD® If you find errors in this book, please report them to Greg Lehey <grog@Free- BSD.org> for inclusion in the errata list. The Complete FreeBSD® Fourth Edition Tenth anniversary version, 24 February 2006 Greg Lehey The Complete FreeBSD® by Greg Lehey <[email protected]> Copyright © 1996, 1997, 1999, 2002, 2003, 2006 by Greg Lehey. This book is licensed under the Creative Commons “Attribution-NonCommercial-ShareAlike 2.5” license. The full text is located at http://creativecommons.org/licenses/by-nc-sa/2.5/legalcode. You are free: • to copy, distribute, display, and perform the work • to make derivative works under the following conditions: • Attribution. You must attribute the work in the manner specified by the author or licensor. • Noncommercial. You may not use this work for commercial purposes. This clause is modified from the original by the provision: You may use this book for commercial purposes if you pay me the sum of USD 20 per copy printed (whether sold or not). You must also agree to allow inspection of printing records and other material necessary to confirm the royalty sums. The purpose of this clause is to make it attractive to negotiate sensible royalties before printing. • Share Alike. If you alter, transform, or build upon this work, you may distribute the resulting work only under a license identical to this one. • For any reuse or distribution, you must make clear to others the license terms of this work. • Any of these conditions can be waived if you get permission from the copyright holder. Your fair use and other rights are in no way affected by the above.
    [Show full text]
  • Iscsi SAN Configuration Guide Update 2 and Later for ESX Server 3.5, ESX Server 3I Version 3.5, Virtualcenter 2.5 Iscsi SAN Configuration Guide
    iSCSI SAN Configuration Guide Update 2 and later for ESX Server 3.5, ESX Server 3i version 3.5, VirtualCenter 2.5 iSCSI SAN Configuration Guide iSCSI SAN Configuration Guide Revision: 20090313 Item: EN-000035-01 You can find the most up-to-date technical documentation on our Web site at: http://www.vmware.com/support/ The VMware Web site also provides the latest product updates. If you have comments about this documentation, submit your feedback to: [email protected] © 2007–2009 VMware, Inc. All rights reserved. This product is protected by U.S. and international copyright and intellectual property laws. VMware products are covered by one or more patents listed at http://www.vmware.com/go/patents. VMware, the VMware “boxes” logo and design, Virtual SMP and VMotion are registered trademarks or trademarks of VMware, Inc. in the United States and/or other jurisdictions. All other marks and names mentioned herein may be trademarks of their respective companies. VMware, Inc. 3401 Hillview Ave. Palo Alto, CA 94304 www.vmware.com 2 VMware, Inc. Contents About This Book 7 1 Using ESX Server with a Storage Area Network 11 Understanding Virtualization 12 Network Virtualization 12 Storage Virtualization 12 Storage Area Network Concepts 15 Ports 16 Multipathing and Path Failover 17 Storage System Types 17 Target Compared to LUN Representations 17 iSCSI Naming Conventions 19 Overview of Using ESX Server with a SAN 20 Benefits of Using ESX Server with a SAN 20 Use Cases 21 Finding SAN Configuration Information 21 Basics of Using SAN Storage Systems with an ESX Server 22 Network Configuration and Authentication 22 Sharing a VMFS Across ESX Servers 22 Metadata Updates 24 Volume Display and Rescan 24 Levels of Indirection 24 Data Access: VMFS or RDM 25 Third‐Party Management Applications 26 Discovery, Authentication, and Access Control 26 Error Correction 27 Understanding VMFS and SAN Storage Choices 27 Choosing Larger or Smaller LUNs 27 Making LUN Decisions 28 Tips for Making LUN Decisions 29 VMware, Inc.
    [Show full text]
  • Oce14000 10Gbe and 40Gbe Converged Network Adapters And
    CONNECTCONNECT - DATA - DATA SHEET SHEET OCe14000OCe14000 10GbE 10Gb andand 40GbE40Gb ConvergedEthernet Network Network Adapters Adapters High Performance Networking for Enterprise Virtualization and the Cloud High Performance Networking for Enterprise Virtualization and the Cloud OneConnectOneConnect OCe14000 OCe14000 Family Family Overview Key benefits As theOverview fourth generation of the Emulex OneConnect product line, the OCe14000 Key benefits As the fourth generation of the Emulex OneConnect® product line, the OCe14000 n Maximizes server hardware ROI with high virtual family of Converged Network Adapters (CNAs) provides high performance 10Gb n Maximizes server hardware ROI with high family of Ethernet network adapters provides high performance 10Gb Ethernet machine density Ethernet (10GbE) and 40GbE connectivity delivering multiple benefits for the virtual machine density (10GbE) and 40GbE connectivity delivering multiple benefits for the enterprise enterprise cloud, including: n Simplifies deployment of secure, scalable multi- cloud, including: tenantn cloudSimplifies infrastructures deployment of secure, scalable multi- n Increasing data center IT agility and scalability through deployment of a secure n Increasing data center IT agility and scalability through deployment of a secure tenant cloud infrastructures multi-tenant cloud n multi-tenant cloud Minimizes TCO through deployment of heterogeneousn Accelerates workloads applications on Converged performance n Driving scalability and flexibility in converged infrastructures
    [Show full text]
  • Utilising Configuration Management Node Data for Network Infrastructure
    Aalto University School of Science Degree Programme in Computer Science and Engineering Lauri Tirkkonen Utilising configuration management node data for network infrastructure man- agement Master's Thesis Espoo, May 4, 2016 Supervisor: Professor Heikki Saikkonen Advisor: Jaakko Kotim¨akiM.Sc. (Tech.) Aalto University School of Science ABSTRACT OF Degree Programme in Computer Science and Engineering MASTER'S THESIS Author: Lauri Tirkkonen Title: Utilising configuration management node data for network infrastructure man- agement Date: May 4, 2016 Pages: 38 Major: Software Technology Code: T-106 Supervisor: Professor Heikki Saikkonen Advisor: Jaakko Kotim¨akiM.Sc. (Tech.) Configuration management software running on nodes solves problems such as configuration drift on the nodes themselves, but the necessary node configura- tion data can also be utilised in managing network infrastructure, for example to reduce configuration errors by facilitating node life cycle management. Many configuration management software systems depend on a working network, but we can utilise the data to create large parts of the network infrastructure config- uration itself using node data from the configuration management system before the nodes themselves are provisioned, as well as remove obsolete configuration as nodes are decommissioned. Keywords: configuration management, network infrastructure configura- tion, life cycle management Language: English 2 Aalto-yliopisto Perustieteiden korkeakoulu DIPLOMITYON¨ Tietotekniikan koulutusohjelma TIIVISTELMA¨ Tekij¨a: Lauri Tirkkonen
    [Show full text]
  • Ubuntu Server Guide Basic Installation Preparing to Install
    Ubuntu Server Guide Welcome to the Ubuntu Server Guide! This site includes information on using Ubuntu Server for the latest LTS release, Ubuntu 20.04 LTS (Focal Fossa). For an offline version as well as versions for previous releases see below. Improving the Documentation If you find any errors or have suggestions for improvements to pages, please use the link at thebottomof each topic titled: “Help improve this document in the forum.” This link will take you to the Server Discourse forum for the specific page you are viewing. There you can share your comments or let us know aboutbugs with any page. PDFs and Previous Releases Below are links to the previous Ubuntu Server release server guides as well as an offline copy of the current version of this site: Ubuntu 20.04 LTS (Focal Fossa): PDF Ubuntu 18.04 LTS (Bionic Beaver): Web and PDF Ubuntu 16.04 LTS (Xenial Xerus): Web and PDF Support There are a couple of different ways that the Ubuntu Server edition is supported: commercial support and community support. The main commercial support (and development funding) is available from Canonical, Ltd. They supply reasonably- priced support contracts on a per desktop or per-server basis. For more information see the Ubuntu Advantage page. Community support is also provided by dedicated individuals and companies that wish to make Ubuntu the best distribution possible. Support is provided through multiple mailing lists, IRC channels, forums, blogs, wikis, etc. The large amount of information available can be overwhelming, but a good search engine query can usually provide an answer to your questions.
    [Show full text]
  • Feature-Rich and Fast SCSI Target with CTL and ZFS
    Feature-rich and fast SCSI target with CTL and ZFS Alexander Motin <[email protected]> iXsystems, Inc. RuBSD'2014 CTL – CAM Target Layer Frontends Backends isp0 FC CAM target device CAM SCSI CAM SIM initiator block zvol CTL core ... file TCP ICL iSCSI iSCSI target ramdisk /dev/zero offload CTL functional improvements ... for intelligent performance ●I/O optimization for storage specifics ●Thin-/Resource-provisioning ●I/O offload CTL functional improvements Basic SCSI disk (CTL in early 2014): ●READ CAPACITY(10) ● Get block size and number of blocks ●READ ● Read range of blocks ●WRITE ● Write range of blocks # diskinfo -v /dev/da0 /dev/da0 512 # sectorsize 107374182400 # mediasize in bytes (100G) 209715200 # mediasize in sectors CTL functional improvements Advanced Format (512e) SCSI disks: ●READ CAPACITY(16) ● Get physical block size and offset # diskinfo -v /dev/da0 /dev/da0 512 # sectorsize 107374182400 # mediasize in bytes (100G) 209715200 # mediasize in sectors 8192 # stripesize 0 # stripeoffset CTL functional improvements Basic thin-provisioned disks: ●Logical Block Provisioning VPD VAAI TP Reporting ● Get Supported UNMAP commands ●Block Limits VPD page ● Get UNMAP block size ● Get UNMAP parameters limitations ●WRITE SAME with UNMAP ● Unmap sequential range of blocks ●UNMAP VAAI Unmap ● Unmap arbitrary list of blocks ●Proper overflow error reporting VAAI TP Stun CTL functional improvements Featured thin-provisioned disk: ●GET LBA STATUS ● Get provisioning status of specific block(s) Windows defrag CTL functional improvements Featured
    [Show full text]