Atm Topology and Protocols
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
What Is It and How We Use It
Infiniband Overview What is it and how we use it What is Infiniband • Infiniband is a contraction of "Infinite Bandwidth" o can keep bundling links so there is no theoretical limit o Target design goal is to always be faster than the PCI bus. • Infiniband should not be the bottleneck. • Credit based flow control o data is never sent if receiver can not guarantee sufficient buffering What is Infiniband • Infiniband is a switched fabric network o low latency o high throughput o failover • Superset of VIA (Virtual Interface Architecture) o Infiniband o RoCE (RDMA over Converged Ethernet) o iWarp (Internet Wide Area RDMA Protocol) What is Infiniband • Serial traffic is split into incoming and outgoing relative to any port • Currently 5 data rates o Single Data Rate (SDR), 2.5Gbps o Double Data Rate (DDR), 5 Gbps o Quadruple Data Rate (QDR), 10 Gbps o Fourteen Data Rate (FDR), 14.0625 Gbps o Enhanced Data Rate (EDR) 25.78125 Gbps • Links can be bonded together, 1x, 4x, 8x and 12x HDR - High Data Rate NDR - Next Data Rate Infiniband Road Map (Infiniband Trade Association) What is Infiniband • SDR, DDR, and QDR use 8B/10B encoding o 10 bits carry 8 bits of data o data rate is 80% of signal rate • FDR and EDR use 64B/66B encoding o 66 bits carry 64 bits of data Signal Rate Latency SDR 200ns DDR 140ns QDR 100ns Hardware 2 Hardware vendors • Mellanox o bought Voltaire • Intel o bought Qlogic Infiniband business unit Need to standardize hardware. Mellanox and Qlogic cards work in different ways. -
SUCCESS STORY. Security Software Porting from Netware to Novell Linux
SUCCESS STORY. >_ >_ Security Software Porting From Netware To Novell >_ Linux About the The Client is a leader in the development of real-time monitoring, auditing and computer forensics Client technologies for Windows and Novell networks. Its ⧄agship product is used to secure the assets of the world’s largest corporations, banks, and government agencies, educational and healthcare institutions. Business The Client had a signi⣴cant install base for its ⧄agship software on Novell Netware OS. Novell’s Challenge adoption of Linux as the migration path for Netware OS created both an opportunity and a challenge for the Client. To continue supporting its Netware user base through Linux migration, the client wanted to port its monitoring and auditing solution to support SUSE Linux, Novell eDirectory and NSS (Novell Storage Services). The Client did not have in-house Linux Systems programming expertise, especially with ⣴le systems & security knowledge and approached Silicus. Silicus The software had two parts – the agent (client) and the server. The agent was a module that was Solution installed on the PC’s to be audited, and sends information on PC activities to a server installed at a centralized location. Termination of Netware required the agent to be re-developed/ported to Linux. Silicus commenced a feasibility study to address a few unknowns in the project: • Identify the auditing modules available on SuSE Linux • XML libraries that could be used • Multi-threading architecture to be used for agent development Silicus created a software architecture and design for the remote management agent. The agent was developed leveraging 3rd party tools to perform the auditing, monitoring of the Linux systems and communication with the remote server. -
FICON Native Implementation and Reference Guide
Front cover FICON Native Implementation and Reference Guide Architecture, terminology, and topology concepts Planning, implemention, and migration guidance Realistic examples and scenarios Bill White JongHak Kim Manfred Lindenau Ken Trowell ibm.com/redbooks International Technical Support Organization FICON Native Implementation and Reference Guide October 2002 SG24-6266-01 Note: Before using this information and the product it supports, read the information in “Notices” on page vii. Second Edition (October 2002) This edition applies to FICON channel adaptors installed and running in FICON native (FC) mode in the IBM zSeries procressors (at hardware driver level 3G) and the IBM 9672 Generation 5 and Generation 6 processors (at hardware driver level 26). © Copyright International Business Machines Corporation 2001, 2002. All rights reserved. Note to U.S. Government Users Restricted Rights -- Use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM Corp. Contents Notices . vii Trademarks . viii Preface . ix The team that wrote this redbook. ix Become a published author . .x Comments welcome. .x Chapter 1. Overview . 1 1.1 How to use this redbook . 2 1.2 Introduction to FICON . 2 1.3 zSeries and S/390 9672 G5/G6 I/O connectivity. 3 1.4 zSeries and S/390 FICON channel benefits . 5 Chapter 2. FICON topology and terminology . 9 2.1 Basic Fibre Channel terminology . 10 2.2 FICON channel topology. 12 2.2.1 Point-to-point configuration . 14 2.2.2 Switched point-to-point configuration . 15 2.2.3 Cascaded FICON Directors configuration. 16 2.3 Access control. 18 2.4 Fibre Channel and FICON terminology. -
Novell Management Tools
04 0789729849_ch03.qxd 11/10/03 12:43 PM Page 91 CHAPTER 3 Novell Management Tools Using ConsoleOne ConsoleOne is a Java-based tool for managing your network and its resources. It can be launched by running CONSOLEONE.EXE from where it was installed (default: SYS:PUBLIC\MGMT\CONSOLEONE\1.2\BIN). By default, it lets you manage Novell eDirectory objects, schema, parti- tions, and replicas and NetWare server resources. If you install other Novell products, the appropriate management capabil- ities are automatically snapped into the version of ConsoleOne installed on that server. ConsoleOne is installed during the NetWare 6.5 installation, but can also be re-installed or installed locally from the Novell client’s CD. ConsoleOne also supports remote server console access through a Java applet called RConsoleJ. To access the NetWare 6.5 server console remotely, launch ConsoleOne and browse to the desired server. Select Tools, and then Remote Console. Accessing Web Manager Web Manager is a Web-based “home page” for accessing most of the NetWare 6.5 Web-based tools and services. To access Web Manager, open your Web browser and enter your Web server’s domain name or IP address, followed by a colon and the Web Manager port, which by default is 2200. For example: 04 0789729849_ch03.qxd 11/10/03 12:43 PM Page 92 92 PART I Getting Started https://www.quills.com:2200 or https://137.65.192.1:2200 Accessing iManager iManager provides role-based management of your NetWare network, together with a nearly comprehensive set of administrative tools. -
Novell Cluster Services,. for Linux. and Netware
Novell Cluster Services,. for Linux. and NetWare. ROB BASTIAANSEN SANDER VAN VUGT Novell PRESS. Novell. Published by Pearson Education, Inc. 800 East 96th Street, Indianapolis, Indiana 46240 USA Table of Contents Introduction 1 CHAPTER 1: Introduction to Clustering and High Availability 5 Novell Cluster Services Defined 5 Shared Disk Access 6 Secondary IP Addresses 7 Clustering Terminology 8 High-Availability Solutions Overview 12 Novell Cluster Services 12 Business Continuity Clustering 13 PolyServe Matrix Server 15 Heartbeat Subsystem for High-Availability Linux 16 When Not to Cluster Applications 16 Availability Defined 18 High Availability Defined 18 Calculating Average Downtime 21 Avoiding Downtime 22 Hardware 22 Environment 23 Software 23 Procedures 24 Novell Cluster Services Requirements 24 Hardware Requirements 24 Software Requirements 26 CHAPTER 2: Examining Novell Cluster Services Architecture 27 Novell Cluster Services Objects and Modules 27 Cluster eDirectory Objects 28 Cluster Modules 31 IH Novell Cluster Services for Linux and NetWare Heartbeats, Epoch Numbers, and the Split Brain Detector 35 Removing a Failing Slave Node 36 Removing a Failed Master Node 37 Summary 37 CHAPTER 3: Clustering Design 39 Cluster Design Guidelines 39 How Many Nodes to Choose 39 Using a Heartbeat LAN or Not 40 Use NIC Teaming 41 Choosing Storage Methods 42 Mirror the Split Brain Detector Partition 48 Selecting Applications to Run in a Cluster 48 eDirectory Cluster Guidelines 50 Creating a Failover Matrix 52 Application-Specific Design Guidelines -
Infiniband Event-Builder Architecture Test-Beds for Full Rate Data
International Conference on Computing in High Energy and Nuclear Physics (CHEP 2010) IOP Publishing Journal of Physics: Conference Series 331 (2011) 022008 doi:10.1088/1742-6596/331/2/022008 Infiniband Event-Builder Architecture Test-beds for Full Rate Data Acquisition in LHCb Enrico Bonaccorsi and Juan Manuel Caicedo Carvajal and Jean-Christophe Garnier and Guoming Liu and Niko Neufeld and Rainer Schwemmer CERN, 1211 Geneva 23, Switzerland E-mail: [email protected] Abstract. The LHCb Data Acquisition system will be upgraded to address the requirements of a 40 MHz readout of the detector. It is not obvious that a simple scale-up of the current system will be able to meet these requirements. In this work we are therefore re-evaluating various architectures and technologies using a uniform test-bed and software framework. Infiniband is a rather uncommon technology in the domain of High Energy Physics data acquisition. It is currently mainly used in cluster based architectures. It has however interesting features which justify our interest : large bandwidth with low latency, a minimal overhead and a rich protocol suite. An InfiniBand test-bed has been and set-up, and the purpose is to have a software interface between the core software of the event-builder and the software related to the communication protocol. This allows us to run the same event-builder over different technologies for comparisons. We will present the test-bed architectures, and the performance of the different entities of the system, sources and destinations, according to their implementation. These results will be compared with 10 Gigabit Ethernet testbed results. -
Test, Simulation, and Integration of Fibre Channel Networked Systems by Mike Glass Principal Marketing Engineer Data Device Corporation
Test, Simulation, and Integration of Fibre Channel Networked Systems By Mike Glass Principal Marketing Engineer Data Device Corporation September 2006 Test, Simulation, and Integration Of Fibre Channel Networked Systems Introduction Fibre Channel is a high-speed networking technology deployed on a number of military/aerospace platforms and programs. These include F- 18E/F, F-16, F-35, B1-B, B-2, E-2D, the Apache Longbow and MMH helicopters, and AESA Radar. Applications for Fibre Channel include mission computers, processor and DSP clusters; data storage; video processing, distribution, and displays; sensors such as radar, FLIR, and video; serial backplanes and IFF. Basic characteristics of Fibre Channel include a choice of copper and optical media options; 1 and 2 Gb operation, including auto-speed negotiation; and operation on multiple topologies including point-to-point, arbitrated loop, and switched fabric. In addition, Fibre Channel provides low latency to the single-digit microsecond level; address scalability with a space up to 224 ports; broadcast and multicast operation; unacknowledged and acknowledged classes of service; a means for providing variable quality of service (QoS); and multiple upper layer protocols (ULPs). Test and simulation applications for deployable Fibre Channel networks include the development of board and box-level systems and sub-systems, network integration, production test, and equipment maintenance. For software development and network integration, it’s often necessary to rely on Fibre Channel testers and analyzers to simulate unavailable equipment for traffic generation and monitoring. Some development and integration environments provide demanding requirements for real-time data monitoring, storage, and subsequent offline analysis. For production test and field maintenance, low cost testers are often better-suited than higher-priced analyzers. -
Etsi Gr Ip6 009 V1.1.1 (2017-03)
ETSI GR IP6 009 V1.1.1 (2017-03) GROUP REPORT IPv6-based Industrial Internet leveraging 6TiSCH technology Disclaimer The present document has been produced and approved by the IPv6 Integration (IP6) ETSI Industry Specification Group (ISG) and represents the views of those members who participated in this ISG. It does not necessarily represent the views of the entire ETSI membership. 2 ETSI GR IP6 009 V1.1.1 (2017-03) Reference DGR/IP6-0009 Keywords 6TiSCH, IPv6, network ETSI 650 Route des Lucioles F-06921 Sophia Antipolis Cedex - FRANCE Tel.: +33 4 92 94 42 00 Fax: +33 4 93 65 47 16 Siret N° 348 623 562 00017 - NAF 742 C Association à but non lucratif enregistrée à la Sous-Préfecture de Grasse (06) N° 7803/88 Important notice The present document can be downloaded from: http://www.etsi.org/standards-search The present document may be made available in electronic versions and/or in print. The content of any electronic and/or print versions of the present document shall not be modified without the prior written authorization of ETSI. In case of any existing or perceived difference in contents between such versions and/or in print, the only prevailing document is the print of the Portable Document Format (PDF) version kept on a specific network drive within ETSI Secretariat. Users of the present document should be aware that the document may be subject to revision or change of status. Information on the current status of this and other ETSI documents is available at https://portal.etsi.org/TB/ETSIDeliverableStatus.aspx If you find errors in the present document, please send your comment to one of the following services: https://portal.etsi.org/People/CommiteeSupportStaff.aspx Copyright Notification No part may be reproduced or utilized in any form or by any means, electronic or mechanical, including photocopying and microfilm except as authorized by written permission of ETSI. -
News Letter V7
Robin MacGillivray, BCS President, SBC West Message to Telecom Consultants Let me start out by wishing each of you a very Happy New Year, with hopes that 2004 will be your most successful year ever! Through your recommendations you have influenced millions of UPDATE dollars in revenue for the SBC family of companies Solutions for Success and it’s my commitment to you to do all we can to continue to earn your trust and confidence in Consultant/Vendor Sales Group 2004. That’s what “Stand & Deliver” is all about. February 2004 It’s a theme we’re using internally to constantly remind ourselves of our promise to you that we will always be the brand you can trust to deliver World-Class products and reliable service to your clients. It’s my personal goal to Stand & Deliver beyond your expectations. We believe the best competitive strategy we have for keeping Douglas Ireland customers is our ability to meet their needs, the first time, on time, SM everytime. There’s probably nothing more critical to the success of SBC FreedomLink Creating Special our business. As we enter a new year, there are a lot of things going National Wi-Fi Hotspot Network on that may directly impact you and your clients. Here are a few: ͷ SBC Internet Services, Inc. is now creating one of the nation’s We’re now offering Long Distance at incredible prices in all 13 states SBC Long Distance serves. largest wireless Internet access networks, commonly known as ͷ We’re offering all sorts of spectacular packages and bundles to Wi-Fi hotspots, under the FreedomLinkSM marquee. -
DR DOS for the Zfx86
DR DOS for the ZFx86 Cost Effective, Reliable, Rapid Deployment of Embedded Systems w DR DOS on the ZFx86 gets products to market quickly at the lowest development and ownership cost. w Thousands of compatible applications, drivers and utilities available free or at minimal cost. w Full documentation available. DR DOS 7.03 Key Features of DR DOS Online Manual - DRDOS is supplied with a complete The ideal embedded DOS system, online manual that contains detailed information about all designed for out-of-the-box of the features of the operating system including the basic implementation into ROM or Flash commands, and the advanced utilities. It also has online ROM with tools and associated help available for all its commands. documents available in the DRDOS Memory Management - Memory management features OEM Documentation Kit. include a set of device drivers and commands that enable w 100% MS-DOS 6.22 compatible.. you to manage memory efficiently and make as much memory as possible available to your applications. w Comprehensive DOS utility set DOS Protected Mode Services - DOS Protected Mode w Multitasking, with API for developers Services (DPMS) interface allows specially-implemented w DPMS memory manager in addition to DPMI device drivers and TSRs to operate in extended memory. w Stacker disk compression This makes more memory within the first megabyte w NWCACHE - disk caching program available to applications and other conventionally-written drivers and TSRs. Both Stacker* (the disk compression w EMM386 memory manager program), and NWCACHE (the disk cache) use DPMS. w DOS Protected Mode Services (DPMS) Disk Compression - The disk compression component w Multitasking enables you to store more information by compressing the w DR-DOS provides a full multitasking environment data. -
Putting Switched Fabric to Work for Software Radio
PUTTING SWITCHED FABRIC TO WORK FOR SOFTWARE RADIO Rodger H. Hosking (Pentek, Inc., Upper Saddle River, NJ, USA, [email protected]) ABSTRACT In order to take advantage of the wealth of high- volume, low-cost devices for mass-market electronics, and The most difficult problem for designers of high- to reap the same benefits of easier connectivity, even the performance, software radio systems is simply moving data most powerful high-end software radio RISC and DSP within the system because of data throughput limitations. processors from Freescale and Texas Instruments are now Driving this dilemma are processors with higher clock rates sporting gigabit serial interfaces. and wider buses, data converter products with higher sampling rates, more complex digital communication 2. GIGABIT SERIAL STANDARDS standards with increased bandwidths, disk storage devices with faster I/O rates, FPGAs and DSPs offering incredible The descriptive phrase “gigabit serial” covers a truly diverse computational rates, and system connections and network range of implementations and application spaces. Figure 1 links operating at higher speeds. shows most of the popular standards used in embedded Traditional system architectures relying on buses and systems suitable for software radio, along with how each parallel connections between system boards and mezzanines standard is normally deployed in a system. fall far short of delivering the required peak rates, and suffer even worse if they must be shared and arbitrated. New Standard Main Application strategies for solving these problems exploit gigabit serial Gigabit Ethernet Computer Networking links and switched fabric standards to create significantly FibreChannel Data Storage more powerful architectures ideally suited for embedded software radio systems. -
Brocade Mainframe Connectivity Solutions
PART 1: BROCADE MAINFRAME CHAPTER 2 CONNECTIVITY SOLUTIONS The modern IBM mainframe, also known as IBM zEnterprise, has a distinguished 50-year history BROCADEMainframe I/O and as the leading platform for reliability, availability, serviceability, and scalability. It has transformed Storage Basics business and delivered innovative, game-changing technology that makes the extraordinary possible, and has improved the way the world works. For over 25 of those years, Brocade, MAINFRAME the leading networking company in the IBM mainframe ecosystem, has provided non-stop The primary purpose of any computing system is to networks for IBM mainframe customers. From parallel channel extension to ESCON, FICON, process data obtained from Input/Output devices. long-distance FCIP connectivity, SNA/IP, and IP connectivity, Brocade has been there with IBM CONNECTIVITY and our mutual customers. Input and Output are terms used to describe the SOLUTIONStransfer of data between devices such as Direct This book, written by leading mainframe industry and technology experts from Brocade, discusses Access Storage Device (DASD) arrays and main mainframe SAN and network technology, best practices, and how to apply this technology in your storage in a mainframe. Input and Output operations mainframe environment. are typically referred to as I/O operations, abbreviated as I/O. The facilities that control I/O operations are collectively referred to as the mainframe’s channel subsystem. This chapter provides a description of the components, functionality, and operations of the channel subsystem, mainframe I/O operations, mainframe storage basics, and the IBM System z FICON qualification process. STEVE GUENDERT DAVE LYTLE FRED SMIT Brocade Bookshelf www.brocade.com/bookshelf i BROCADE MAINFRAME CONNECTIVITY SOLUTIONS STEVE GUENDERT DAVE LYTLE FRED SMIT BROCADE MAINFRAME CONNECTIVITY SOLUTIONS ii © 2014 Brocade Communications Systems, Inc.