Intel® Omni-Path Fabric Host Software — User Guide
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Rapidio® Connections
® RapidIO Connections Q4 2011 Forward: RapidIO has had an amazing year with many new product announcements, the launch of the RapidIO 10xN roadmap--with serial lanes up to 10 Gbps and port data rates that go up to 160 Gbps--and production shipments of RapidIO Gen2 based systems. Don’t miss this issue’s links to many of the recent news releases and published articles. RapidIO is seeing excitement beyond its traditional wireless and defense/aerospace markets with high interest in the Cloud Computing market, both in the data center and in server systems being pushed to the base station in wireless. The introduction of new Gen 2 RapidIO products, and a number of single board computers with RapidIO and the Intel® Core i7 processor family. -Tom Cox. In This Issue Association News RapidIO Trade Association Releases Roadmap to 100 Gbps+ per port with RapidIO® 10xN Press Release - July 2011 Pushing the boundaries of performance for peer-to-peer embedded processing for the wireless, server, imaging, medical, industrial, defense and aerospace markets, the RTA announced its roadmap beyond RapidIO Gen2 in July of 2011. The new spec is S-RIO 10xN and it delivers 10 Gbps per serial lane and beyond. The S-RIO 10xNwill actually scale to speeds per serial lane up to 25 Gbps in the next revision of the standard. In x4 configuration, it means 40-100 Gbps per port. http://www.rapidio.org/news/pr/view?item_key=6ff6de5e4d4d09ba26a7242c34b85c28c9dd4af6 Industry Insights The OpenFabrics Alliance and OFED By OpenFabrics Alliance The OpenFabrics Alliance (OFA) is an open source community with a charter to develop, test and license high performance networking software for servers and storage systems. -
Storage Administration Guide Storage Administration Guide SUSE Linux Enterprise Server 12 SP4
SUSE Linux Enterprise Server 12 SP4 Storage Administration Guide Storage Administration Guide SUSE Linux Enterprise Server 12 SP4 Provides information about how to manage storage devices on a SUSE Linux Enterprise Server. Publication Date: September 24, 2021 SUSE LLC 1800 South Novell Place Provo, UT 84606 USA https://documentation.suse.com Copyright © 2006– 2021 SUSE LLC and contributors. All rights reserved. Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or (at your option) version 1.3; with the Invariant Section being this copyright notice and license. A copy of the license version 1.2 is included in the section entitled “GNU Free Documentation License”. For SUSE trademarks, see https://www.suse.com/company/legal/ . All other third-party trademarks are the property of their respective owners. Trademark symbols (®, ™ etc.) denote trademarks of SUSE and its aliates. Asterisks (*) denote third-party trademarks. All information found in this book has been compiled with utmost attention to detail. However, this does not guarantee complete accuracy. Neither SUSE LLC, its aliates, the authors nor the translators shall be held liable for possible errors or the consequences thereof. Contents About This Guide xii 1 Available Documentation xii 2 Giving Feedback xiv 3 Documentation Conventions xiv 4 Product Life Cycle and Support xvi Support Statement for SUSE Linux Enterprise Server xvii • Technology Previews xviii I FILE SYSTEMS AND MOUNTING 1 1 Overview -
Mac Labs. More Creativity
Mac labs. More creativity. More engagement. More learning. Every Mac comes ready for 21st-century learning. Mac OS X Server creates the ultimate lab environment. Creativity and innovation are critical skills in the modern Mac OS X Server includes simplified tools for creating wikis and workplace, which makes them critical skills to teach today’s blogs, so students and staff can set up and manage them with students. Apple offers comprehensive lab solutions for your little or no help from IT. The Spotlight Server feature makes it school or university that will creatively engage students and easy for students collaborating on projects to find and view motivate them to take an interest in their own learning. The content stored anywhere on the network. Of course, access iLife ’09 suite of applications comes standard on every Mac, controls are built in, so they only get the search results they’re so students can immediately begin creating podcasts, editing authorised to see. Mac OS X Server also allows for real-time videos, composing music, producing photo essays and more. streaming of audio and video content, which lets everyone get right to work instead of waiting for downloads – saving major Mac or PC? Have both. storage across your network. Now one lab really can solve all your computing needs. Every Mac is powered by an Intel processor and features Launch careers from your Mac lab. Mac OS X – the world’s most advanced operating system. As an Apple Authorised Training Centre for Education, your Mac OS X comes with an amazing dual-boot feature called school can build a bridge between students and real-world Boot Camp that lets students run Windows XP or Vista natively careers. -
Nvmf Based Integration of Non-Volatile Memory in A
14th ANNUAL WORKSHOP 2018 NVMf based Integration of Non-volatile Memory in a Distributed System - Lessons learned Jonas Pfefferle, Bernard Metzler, Patrick Stuedi, Animesh Trivedi and Adrian Schuepbach IBM Zurich Research April 2018 Outline . Target Distributed System: Apache Crail . NVMf • Basic operation • Available implementations . The Crail NVMf Tier • Integration with Crail • Implementation details • A new NVMf host library . Measurements • Microbenchmarks • TeraSort • TPC-DS workload . Summary & Outlook 2 OpenFabrics Alliance Workshop 2018 The Target System: Apache Crail (Incubating) . Targets I/O acceleration for Big Data frameworks due to user level and asynchronous I/O . Implements efficient distributed store for ephemeral data at native hardware speeds . Unifies isolated, incompatible efforts to integrate high performance I/O with BigData processing frameworks Data Processing Framework . Enables dramatically faster job (Spark, TensorFlow, λ Compute, …) completion 10th of GB/s . Applicable to Spark, Flink, < 10 µsec FS Streaming KV HDFS TensorFlow, Caffee, … Apache Crail . Flexibly makes best use of available I/O infrastructure TCP RDMA NVMf SPDK . Apache Incubator Project Fast Network (100Gb/s RoCE, etc.) (since 11/17) DRAM NVMe PCM … GPU 3 OpenFabrics Alliance Workshop 2018 NVMf in a Nutshell . Extends low latency NVMe interface over distance NVM read command • Retains BW and latency (< 5us additional delay) • Allows to retain user level I/O capability NVMf RDMA RDMA NVMf NVMe SQ/CQ QP • Message based NVMe operations QP SQ/CQ Device • Host/Target model IO cmd . Defined transports: RDMA (IB, RoCE, post_send iWarp), or FC SEND • Mapping of I/O Submission + Completion Queue recv comp to RDMA QP model NVMe read . Fabrics + Admin Commands • Discover, Connect, Authenticate, … CMD comp . -
For Immediate Release
an ellisys company FOR IMMEDIATE RELEASE SerialTek Contact: Simon Thomas, Director, Sales and Marketing Phone: +1-720-204-2140 Email: [email protected] SerialTek Debuts PCIe x16 Gen5 Protocol Analysis System and Web Application New Kodiak™ Platform Brings SerialTek Advantages to More Computing and Data Storage Markets Longmont, CO, USA — February 24, 2021 — SerialTek, a leading provider of protocol test solutions for PCI Express®, NVM Express®, Serial Attached SCSI, and Serial ATA, today introduced an advancement in the PCIe® test and analysis market with the release of the Kodiak PCIe x16 Gen5 Analysis System, as well as the industry’s first calibration-free PCIe x16 add-in-card (AIC) interposer and a new web-based BusXpert™ user interface to manage and analyze traces more efficiently than ever. The addition of PCIe x16 Gen5 to the Kodiak analyzer and SI-Fi™ interposer family brings previously unavailable analysis capabilities and efficiencies to computing, datacenter, networking, storage, AI, and other PCIe x16 Gen5 applications. With SerialTek’s proven calibration-free SI-Fi interposer technology, the Kodiak’s innovative state-of-the-art design, and the new BusXpert analyzer software, users can more easily set up the analyzer hardware, more accurately probe PCIe signals, and more efficiently analyze traces. Kodiak PCIe x16 Gen5 Analysis System At the core of the Kodiak PCIe x16 analyzer is an embedded hardware architecture that delivers substantial and unparalleled advancements in capture, search, and processing acceleration. Interface responsiveness is markedly advanced, searches involving massive amounts of data are fast, and hardware filtering is flexible and powerful. “Once installed in a customer test environment the Kodiak’s features and benefits are immediately obvious,” said Paul Mutschler, CEO of SerialTek. -
Integrated Step-Down Converter with SVID Interface for Intel® CPU
Product Order Technical Tools & Support & Folder Now Documents Software Community TPS53820 SLUSE33 –FEBRUARY 2020 Integrated Step-Down Converter with SVID Interface for Intel® CPU Power Supply 1 Features 3 Description The TPS53820 device is D-CAP+ mode integrated 1• Single chip to support Intel VR13.HC SVID POL applications step-down converter for low current SVID rails of Intel CPU power supply. It provides up to two outputs to • Two outputs to support VCCANA (5.5 A) and power the low current SVID rails such as VCCANA P1V8 (4 A) (5.5 A) and P1V8 (4 A). The device employs D-CAP+ • D-CAP+™ Control for Fast Transient Response mode control to provide fast load transient • Wide Input Voltage (4.5 V to 15 V) performance. Internal compensation allows ease of use and reduces external components. • Differential Remote Sense • Programmable Internal Loop Compensation The device also provides telemetry, including input voltage, output voltage, output current and • Per-Phase Cycle-by-Cycle Current Limit temperature reporting. Over voltage, over current and • Programmable Frequency from 800 kHz to 2 MHz over temperature protections are provided as well. 2 • I C System Interface for Telemetry of Voltage, The TPS53820 device is packaged in a thermally Current, Output Power, Temperature, and Fault enhanced 35-pin QFN and operates between –40°C Conditions and 125°C. • Over-Current, Over-Voltage, Over-Temperature (1) protections Table 1. Device Information • Low Quiescent Current PART NUMBER PACKAGE BODY SIZE (NOM) • 5 mm × 5 mm, 35-Pin QFN, PowerPAD Package TPS53820 RWZ (35) 5 mm × 5 mm (1) For all available packages, see the orderable addendum at 2 Applications the end of the data sheet. -
Introduction to Intel® FPGA IP Cores
Introduction to Intel® FPGA IP Cores Updated for Intel® Quartus® Prime Design Suite: 20.3 Subscribe UG-01056 | 2020.11.09 Send Feedback Latest document on the web: PDF | HTML Contents Contents 1. Introduction to Intel® FPGA IP Cores..............................................................................3 1.1. IP Catalog and Parameter Editor.............................................................................. 4 1.1.1. The Parameter Editor................................................................................. 5 1.2. Installing and Licensing Intel FPGA IP Cores.............................................................. 5 1.2.1. Intel FPGA IP Evaluation Mode.....................................................................6 1.2.2. Checking the IP License Status.................................................................... 8 1.2.3. Intel FPGA IP Versioning............................................................................. 9 1.2.4. Adding IP to IP Catalog...............................................................................9 1.3. Best Practices for Intel FPGA IP..............................................................................10 1.4. IP General Settings.............................................................................................. 11 1.5. Generating IP Cores (Intel Quartus Prime Pro Edition)...............................................12 1.5.1. IP Core Generation Output (Intel Quartus Prime Pro Edition)..........................13 1.5.2. Scripting IP Core Generation.................................................................... -
ISS 2016 Bootcamp: Intel Omni-Path Architecture
Intel® Scalable System Framework A Configurable Design Philosophy Extensible to a Wide Range of Workloads Small Clusters Through Supercomputers Compute Memory/Storage Compute and Data-Centric Computing Fabric Software Standards-Based Programmability Intel Silicon On-Premise and Cloud-Based Photonics Intel® Xeon® Processors Intel® Solutions for Lustre* Intel® Omni-Path Architecture HPC System Software Stack Intel® Xeon Phi™ Processors Intel® SSDs Intel® True Scale Fabric Intel® Software Tools Intel® Xeon Phi™ Coprocessors Intel® Optane™ Technology Intel® Ethernet Intel® Cluster Ready Program Intel® Server Boards and Platforms 3D XPoint™ Technology Intel® Silicon Photonics Intel® Visualization Toolkit 1 YANG YANGUO May 2016 Intel Confidential Agenda Quick Overview: HPC Fabrics What is Intel® 100Gb Omni-Path Architecture(OPA)? Why is Intel 100Gb OPA Summary Intel® Solutions Summit 2016 Intel Confidential 3 Intel Confidential What is Different Between Networks and Fabrics? Network: Universal interconnect designed to HPC Fabric: Optimized interconnect allows allow any-and-all systems to communicate many nodes to perform as a single system Intel® Omni-Path Architecture or Infiniband Key NETWORK (Ethernet) Attributes: Key FABRIC Attributes: . Flexibility for any application . Targeted for specific applications . Designed for universal communication . Optimized for performance and efficiency . Extensible configuration . Engineered topologies . Multi-vendor components . Single-vendor solutions Intel® Solutions Summit 2016 Intel Confidential 5 Fabric: InfiniBand* and OPA InfiniBand/OPA is a multi-lane, high-speed serial interconnect (Copper or Fiber) . Typically presented as a 4x solution . Speeds: 40Gb/s (M & Intel QDR), 56Gb/s (M FDR), 100Gb/s (EDR & Intel OPA) High bandwidth, low latency HPC interconnect for commodity servers . Ethernet switch latency is typically measured in μs, but InfiniBand/OPA is in nanoseconds . -
Introducing the New 11Th Gen Intel® Core™ Desktop Processors
Product Brief 11th Gen Intel® Core™ Desktop Processors Introducing the New 11th Gen Intel® Core™ Desktop Processors The 11th Gen Intel® Core™ desktop processor family puts you in control of your compute experience. It features an innovative new architecture for reimag- ined performance, immersive display and graphics for incredible visuals, and a range of options and technologies for enhanced tuning. When these advances come together, you have everything you need for fast-paced professional work, elite gaming, inspired creativity, and extreme tuning. The 11th Gen Intel® Core™ desktop processor family gives you the power to perform, compete, excel, and power your greatest contributions. Product Brief 11th Gen Intel® Core™ Desktop Processors PERFORMANCE Reimagined Performance 11th Gen Intel® Core™ desktop processors are intelligently engineered to push the boundaries of performance. The new processor core architecture transforms hardware and software efficiency and takes advantage of Intel® Deep Learning Boost to accelerate AI performance. Key platform improvements include memory support up to DDR4-3200, up to 20 CPU PCIe 4.0 lanes,1 integrated USB 3.2 Gen 2x2 (20G), and Intel® Optane™ memory H20 with SSD support.2 Together, these technologies bring the power and the intelligence you need to supercharge productivity, stay in the creative flow, and game at the highest level. Product Brief 11th Gen Intel® Core™ Desktop Processors Experience rich, stunning, seamless visuals with the high-performance graphics on 11th Gen Intel® Core™ desktop -
Proposal Regarding USENIX and Openfabrics Alliance Ellie Young, 9/18/2008
Proposal Regarding USENIX and OpenFabrics Alliance Ellie Young, 9/18/2008 For the past two months a group from the OpenFabrics Alliance and USENIX management have been working together to reach an agreement on how best to explore the feasibility and desirability of OFA and USENIX combining forces in a way that would work to the advantage of both organizations. Now the USENIX Board must decide whether we should take the next step: sending a MOU to provide services to OFA as a 501(c)(6) under our 501(c)(3). We realized last week that in order for us to announce this at the obvious juncture, the upcoming Supercomputing Conference in November, we need to have a decision from OFA by September 29. In the past month USENIX has been given a consulting membership in OFA. Some of us have been participating in marketing, legal, and executive working group meetings. I think we have now spent enough time "getting to know one another." Background Last year, in response to the Board's request that we reach out to open source software projects and look beyond the academic and research community we currently serve, Clem Cole introduced us to several members of the OpenFabrics Alliance. We made a proposal to organize their annual Sonoma Workshop this past spring, but they decided to continue organizing this year’s event on their own. We were contacted again this summer. At a meeting in August, part of which Matt Blaze attended, we discussed the current state of OFA. Our notes on that meeting include: Software development: Development is the most valuable activity for OFA, but it's on its own track (Sonoma workshops, as well as workshops alongside Supercomputing, although there were none this year) and goes pretty smoothly for two software releases per year. -
NVM Express Over Fabrics
NVM Express Over Fabrics Dave Minturn, Intel Corp. OFADevWorkshop NVMe in Fabric Environments • A primary use case for NVMe PCIe SSDs is in an all flash appliance • Hundreds or more SSDs may be attached – too many for PCIe based attach scale-out • Concern: Today remote SSD scale-out over a fabric attach uses SCSI based protocols: Requiring protocol translation Desire best performance and latency from NVMe SSD investment over fabrics like Ethernet with RDMA (iWARP,RoCE), InfiniBand™, and Intel® Omni-Path Architecture Realizing Benefit of Next Gen NVM over Fabrics • PCIe NVMe SSD latency may be < 10 µs with Next Generation NVM • Using a SCSI-based protocol for remote NVMe access ads over 100 µs* in latency • Usage models require efficient write mirroring of PCIe Next Gen NVMe SSDs over fabrics *Source: Intel measurements. Concern: Low latency of Next Gen NVM lost in (SCSI) translation. Why NVMe over Fabrics? Simplicity, Efficiency and End-to-End NVMe Model • Simplicity of protocol enables hardware automated I/O Queues – NVMe transport bridge • No translation to or from another protocol like SCSI (in firmware/software) • Inherent parallelism of NVMe multiple I/O Queues is exposed to the host • NVMe commands and structures are transferred end-to-end • Maintains the NVMe architecture across a range of fabric types • Maintains architecture and software consistency between fabric types by standardizing a common abstraction and encapsulation definition Performance Goal: Make remote NVMe access over fabrics equivalent to local PCIe attached NVMe, -
THE STORAGE PERFORMANCE DEVELOPMENT KIT and NVME-OF Paul Luse Intel Corporation
14th ANNUAL WORKSHOP 2018 THE STORAGE PERFORMANCE DEVELOPMENT KIT AND NVME-OF Paul Luse Intel Corporation Apr 2018 AGENDA Storage • What is SPDK? Performance • The SPDK Community Development Kit • Why are so many storage companies using it? • How is it being used? 2 OpenFabrics Alliance Workshop 2018 WHAT IS SPDK? 3 OpenFabrics Alliance Workshop 2018 WHAT IS SPDK? Scalable and Efficient Software Ingredients Storage . User space, lockless, polled-mode components Performance . Up to millions of IOPS per core Development Kit . Designed to extract maximum performance from non- volatile media Storage Reference Architecture . Optimized for latest generation CPUs and SSDs . Open source composable building blocks (BSD licensed) http://SPDK.IO 4 OpenFabrics Alliance Workshop 2018 SPDK ARCHITECTURE 5 OpenFabrics Alliance Workshop 2018 THE SPDK COMMUNITY 6 OpenFabrics Alliance Workshop 2018 THE SPDK COMMUNITY Full Transparency 7 OpenFabrics Alliance Workshop 2018 WHY ARE SO MANY STORAGE COMPANIES USING SPDK? 8 OpenFabrics Alliance Workshop 2018 SOFTWARE IS BECOMING THE BOTTLENECK SPDK Unlocks New Media Potential 9 OpenFabrics Alliance Workshop 2018 SPDK BENEFITS . Up to 10X MORE IOPS/core for NVMe-oF* vs kernel Storage . Up to 8X MORE IOPS/core for NVMe vs kernel Performance . Up to 50% BETTER tail latency w/RocksDB workloads Development Kit . FASTER TTM w/less RESOURCES vs from scratch . Provides FUTURE PROOFING as technologies evolve Software and workloads used in performance tests may have been optimized for performance only on Intel microprocessors. Performance tests, such as SYSmark and MobileMark, are measured using specific computer systems, components, software, operations and functions. Any change to any of those factors may cause the results to vary.