Administering IBM Platform

Total Page:16

File Type:pdf, Size:1020Kb

Administering IBM Platform Platform LSF Version 9 Release 1.1 Administering Platform LSF SC27-5302-01 Platform LSF Version 9 Release 1.1 Administering Platform LSF SC27-5302-01 Note Before using this information and the product it supports, read the information in “Notices” on page 751. First edition This edition applies to version 9, release 1 of IBM Platform LSF (product number 5725G82) and to all subsequent releases and modifications until otherwise indicated in new editions. © Copyright IBM Corporation 1992, 2013. US Government Users Restricted Rights – Use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM Corp. Contents Managing Your Cluster ........1 Resource Preemption ...........378 Working with Your Cluster .........1 Guaranteed Resource Pools.........383 LSF Daemon Startup Control ........22 Goal-Oriented SLA-Driven Scheduling .....393 Working with Hosts ...........30 Exclusive Scheduling ...........412 Managing Jobs .............66 Working with Queues ...........90 Job Scheduling and Dispatch ....415 LSF Resources .............99 Working with Application Profiles ......415 External Load Indices ..........122 Job Directories and Data..........430 Managing Users and User Groups ......135 Resource Allocation Limits .........433 External Host and User Groups .......140 Reserving Resources ...........445 Between-Host User Account Mapping .....144 Job Dependency and Job Priority .......459 Cross-Cluster User Account Mapping .....150 Job Requeue and Job Rerun ........477 UNIX/Windows User Account Mapping ....154 Job Migration .............482 Job Checkpoint and Restart.........491 Cluster Version Management and Resizable Jobs .............505 Patching on UNIX and Linux .....163 Chunk Jobs and Job Arrays.........514 Scope ................163 Job Packs...............527 Enable LSF HPC Features .........173 Running Parallel Jobs...........529 Monitoring Your Cluster .......179 Job Execution and Interactive Jobs 613 Achieving Performance and Scalability .....179 Runtime Resource Usage Limits .......613 Event Generation ............190 Load Thresholds ............627 Tuning the Cluster............192 Pre-Execution and Post-Execution Processing . 631 Authentication and Authorization ......203 Job Starters ..............646 Submitting Jobs with SSH .........210 Job Controls ..............651 External Authentication ..........214 External Job Submission and Execution Controls 657 Job Email and Job File Spooling .......225 Interactive Jobs with bsub .........674 Non-Shared File Systems .........230 Interactive and Remote Tasks ........685 Error and Event Logging .........235 Troubleshooting and Error Messages .....244 Appendices ............693 Submitting Jobs Using JSDL ........693 Time-Based Configuration ......261 Using lstch ..............703 Time Configuration ...........261 Using Session Scheduler ..........711 Advance Reservation ...........268 Using lsmake .............726 Managing LSF on EGO ..........732 Job Scheduling Policies.......289 Notices ..............751 Preemptive Scheduling ..........289 Specifying Resource Requirements ......304 Trademarks ..............753 Fairshare Scheduling ...........345 © Copyright IBM Corp. 1992, 2013 iii iv Administering IBM Platform LSF Managing Your Cluster “Working with Your Cluster” “LSF Daemon Startup Control” on page 22 “Working with Hosts” on page 30 “Managing Jobs” on page 66 “Working with Queues” on page 90 “LSF Resources” on page 99 “External Load Indices” on page 122 “Managing Users and User Groups” on page 135 “External Host and User Groups” on page 140 “Between-Host User Account Mapping” on page 144 “Cross-Cluster User Account Mapping” on page 150 “UNIX/Windows User Account Mapping” on page 154 Working with Your Cluster “Learn about LSF” “View cluster information” on page 4 “Example directory structures” on page 8 “Add cluster administrators” on page 10 “Control daemons” on page 11 “Control mbatchd” on page 12 “Reconfigure your cluster” on page 14 “Live reconfiguration” on page 16 Learn about LSF Before using LSF for the first time, you should download and read LSF Foundations Guide for an overall understanding of how LSF works. Basic concepts Job states: LSF jobs have the following states: v PEND: Waiting in a queue for scheduling and dispatch v RUN: Dispatched to a host and running v DONE: Finished normally with zero exit value v EXIT: Finished with non-zero exit value v PSUSP: Suspended while pending v USUSP: Suspended by user v SSUSP: Suspended by the LSF system v POST_DONE: Post-processing completed without errors v POST_ERR: Post-processing completed with errors v UNKWN: mbatchd lost contact with sbatchd on the host on which the job runs v WAIT: For jobs submitted to a chunk job queue, members of a chunk job that are waiting to run © Copyright IBM Corp. 1992, 2013 1 v ZOMBI: A job becomes ZOMBI if the execution host is unreachable when a non-rerunnable job is killed or a rerunnable job is requeued Host: An individual computer in the cluster. Each host might have more than one processor. Multiprocessor hosts are used to run parallel jobs. A multiprocessor host with a single process queue is considered a single machine, while a box full of processors that each have their own process queue is treated as a group of separate machines. Tip: The names of your hosts should be unique. They should not be the same as the cluster name or any queue defined for the cluster. Job: A unit of work that is run in the LSF system. A job is a command submitted to LSF for execution, using the bsub command. LSF schedules, controls, and tracks the job according to configured policies. Jobs can be complex problems, simulation scenarios, extensive calculations, anything that needs compute power. Job files When a job is submitted to a queue, LSF holds it in a job file until conditions are right for it to be executed. Then the job file is used to execute the job. UNIX: The job file is a Bourne shell script that is run at execution time. Windows: The job file is a batch file that is processed at execution time. Interactive batch job: A batch job that allows you to interact with the application and still take advantage of LSF scheduling policies and fault tolerance. All input and output are through the terminal that you used to type the job submission command. When you submit an interactive job, a message is displayed while the job is awaiting scheduling. A new job cannot be submitted until the interactive job is completed or terminated. Interactive task: A command that is not submitted to a batch queue and scheduled by LSF, but is dispatched immediately. LSF locates the resources needed by the task and chooses the best host among the candidate hosts that has the required resources and is lightly loaded. Each command can be a single process, or it can be a group of cooperating processes. Tasks are run without using the batch processing features of LSF but still with the advantage of resource requirements and selection of the best host to run the task based on load. Local task: An application or command that does not make sense to run remotely. For example, the ls command on UNIX. 2 Administering IBM Platform LSF Remote task: An application or command that can be run on another machine in the cluster. Host types and host models: Hosts in LSF are characterized by host type and host model. The following example is a host with type X86_64, with host models Opteron240, Opteron840, Intel_EM64T, etc. Host type: The combination of operating system and host CPU architecture. All computers that run the same operating system on the same computer architecture are of the same type - in other words, binary-compatible with each other. Each host type usually requires a different set of LSF binary files. Host model: The host type of the computer, which determines the CPU speed scaling factor that is applied in load and placement calculations. The CPU factor is considered when jobs are being dispatched. Resources: Resource usage: The LSF system uses built-in and configured resources to track resource availability and usage. Jobs are scheduled according to the resources available on individual hosts. Jobs submitted through the LSF system will have the resources they use monitored while they are running. This information is used to enforce resource limits and load thresholds as well as fairshare scheduling. LSF collects information such as: v Total CPU time that is consumed by all processes in the job v Total resident memory usage in KB of all currently running processes in a job v Total virtual memory usage in KB of all currently running processes in a job v Currently active process group ID in a job v Currently active processes in a job On UNIX, job-level resource usage is collected through PIM. Load indices: Load indices measure the availability of dynamic, non-shared resources on hosts in the cluster. Load indices built into the LIM are updated at fixed time intervals. Managing Your Cluster 3 External load indices: Defined and configured by the LSF administrator and collected by an External Load Information Manager (ELIM) program. The ELIM also updates LIM when new values are received. Static resources: Built-in resources that represent host information that does not change over time, such as the maximum RAM available to user processes or the number of processors in a machine. Most static resources are determined by the LIM at start-up time. Static resources can be used to select appropriate hosts for particular jobs that are based on binary architecture, relative CPU speed, and system configuration. Load thresholds: Two types of load thresholds can be configured by your LSF administrator to schedule jobs in queues. Each load threshold specifies a load index value: v loadSched determines the load condition for dispatching pending jobs. If a host’s load is beyond any defined loadSched, a job cannot be started on the host. This threshold is also used as the condition for resuming suspended jobs. v loadStop determines when running jobs should be suspended. To schedule a job on a host, the load levels on that host must satisfy both the thresholds configured for that host and the thresholds for the queue from which the job is being dispatched.
Recommended publications
  • Platform LSF Foundations Chapter 1
    Platform LSF Version 9 Release 1.3 Foundations SC27-5304-03 Platform LSF Version 9 Release 1.3 Foundations SC27-5304-03 Note Before using this information and the product it supports, read the information in “Notices” on page 21. First edition This edition applies to version 9, release 1 of IBM Platform LSF (product number 5725G82) and to all subsequent releases and modifications until otherwise indicated in new editions. Significant changes or additions to the text and illustrations are indicated by a vertical line (|) to the left of the change. If you find an error in any Platform Computing documentation, or you have a suggestion for improving it, please let us know. In the IBM Knowledge Center, add your comments and feedback to any topic. You can also send your suggestions, comments and questions to the following email address: [email protected] Be sure include the publication title and order number, and, if applicable, the specific location of the information about which you have comments (for example, a page number or a browser URL). When you send information to IBM, you grant IBM a nonexclusive right to use or distribute the information in any way it believes appropriate without incurring any obligation to you. © Copyright IBM Corporation 1992, 2014. US Government Users Restricted Rights – Use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM Corp. Contents Chapter 1. IBM Platform LSF: An Job submission .............12 Overview ..............1 Job scheduling and dispatch .........13 Introduction to IBM Platform LSF .......1 Host selection .............15 LSF cluster components...........2 Job execution environment .........15 Chapter 2.
    [Show full text]
  • IBM Platform Computing Cloud Service Ready to Use Platform LSF & Symphony Clusters in the Softlayer Cloud
    Platform Computing IBM Platform Computing Cloud Service Ready to use Platform LSF & Symphony clusters in the SoftLayer cloud February 25, 2014 1 © 2014 IBM Corporation Platform Computing Agenda v Mapping clients needs to cloud technologies v Addressing your pain points v Introducing IBM Platform Computing Cloud Service v Product features and benefits v Use cases v Performance benchmarks 2 © 2014 IBM Corporation Platform Computing HPC cloud characteristics and economics are different than general-purpose computing • High-end hardware and special purpose devices (e.g. GPUs) are typically used to supply the needed processing, memory, network, and storage capabilities • The performance requirements of technical computing and service-oriented workloads means that performance may be impacted in a virtualized cloud environment, especially when latency or I/O is a constraint • HPC cluster/grid utilization is usually in the 70-90% range, removing a major potential advantage of a public cloud service provider for stable workload volumes HPC Workloads Recommended for Private Cloud HPC Workloads with Best Potential for Virtualized Public & Hybrid Cloud Primary HPC Workloads 3 © 2014 IBM Corporation Platform Computing IBM’s HPC cloud strategy provides a flexible approach to address a variety of client needs Private Hybrid Public Clouds Clouds Clouds Evolve existing infrastructure to Enable integrated HPC Cloud to enhance approach to improve Access additional responsiveness, HPC cost and HPC capacity with flexibility, and capability variable cost model
    [Show full text]
  • QEMU Version 4.2.0 User Documentation I
    QEMU version 4.2.0 User Documentation i Table of Contents 1 Introduction ::::::::::::::::::::::::::::::::::::: 1 1.1 Features :::::::::::::::::::::::::::::::::::::::::::::::::::::::: 1 2 QEMU PC System emulator ::::::::::::::::::: 2 2.1 Introduction :::::::::::::::::::::::::::::::::::::::::::::::::::: 2 2.2 Quick Start::::::::::::::::::::::::::::::::::::::::::::::::::::: 2 2.3 Invocation :::::::::::::::::::::::::::::::::::::::::::::::::::::: 3 2.3.1 Standard options :::::::::::::::::::::::::::::::::::::::::: 3 2.3.2 Block device options :::::::::::::::::::::::::::::::::::::: 12 2.3.3 USB options:::::::::::::::::::::::::::::::::::::::::::::: 23 2.3.4 Display options ::::::::::::::::::::::::::::::::::::::::::: 23 2.3.5 i386 target only::::::::::::::::::::::::::::::::::::::::::: 30 2.3.6 Network options :::::::::::::::::::::::::::::::::::::::::: 31 2.3.7 Character device options:::::::::::::::::::::::::::::::::: 38 2.3.8 Bluetooth(R) options ::::::::::::::::::::::::::::::::::::: 42 2.3.9 TPM device options :::::::::::::::::::::::::::::::::::::: 43 2.3.10 Linux/Multiboot boot specific ::::::::::::::::::::::::::: 44 2.3.11 Debug/Expert options ::::::::::::::::::::::::::::::::::: 45 2.3.12 Generic object creation :::::::::::::::::::::::::::::::::: 54 2.3.13 Device URL Syntax ::::::::::::::::::::::::::::::::::::: 66 2.4 Keys in the graphical frontends :::::::::::::::::::::::::::::::: 69 2.5 Keys in the character backend multiplexer ::::::::::::::::::::: 69 2.6 QEMU Monitor ::::::::::::::::::::::::::::::::::::::::::::::: 70 2.6.1 Commands :::::::::::::::::::::::::::::::::::::::::::::::
    [Show full text]
  • Platform LSF Quick Reference IBM Platform LSF 9.1.2 Quick Reference
    Platform LSF Version 9 Release 1.2 Quick Reference GC27-5309-02 Platform LSF Version 9 Release 1.2 Quick Reference GC27-5309-02 Note Before using this information and the product it supports, read the information in “Notices” on page 9. First edition This edition applies to version 9, release 1 of IBM Platform LSF (product number 5725G82) and to all subsequent releases and modifications until otherwise indicated in new editions. Significant changes or additions to the text and illustrations are indicated by a vertical line (|) to the left of the change. If you find an error in any Platform Computing documentation, or you have a suggestion for improving it, please let us know. Send your suggestions, comments and questions to the following email address: [email protected] Be sure include the publication title and order number, and, if applicable, the specific location of the information about which you have comments (for example, a page number or a browser URL). When you send information to IBM, you grant IBM a nonexclusive right to use or distribute the information in any way it believes appropriate without incurring any obligation to you. © Copyright IBM Corporation 1992, 2013. US Government Users Restricted Rights – Use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM Corp. Contents Notices ...............9 Privacy policy considerations ........11 Trademarks ..............11 © Copyright IBM Corp. 1992, 2013 iii iv Platform LSF Quick Reference IBM Platform LSF 9.1.2 Quick Reference Sample UNIX installation directories Daemon error log files Daemon error log files are stored in the directory defined by LSF_LOGDIR in lsf.conf.
    [Show full text]
  • How to Configure the NFX250
    How to Configure the NFX250 Published 2021-07-12 ii Juniper Networks, Inc. 1133 Innovation Way Sunnyvale, California 94089 USA 408-745-2000 www.juniper.net Juniper Networks, the Juniper Networks logo, Juniper, and Junos are registered trademarks of Juniper Networks, Inc. in the United States and other countries. All other trademarks, service marks, registered marks, or registered service marks are the property of their respective owners. Juniper Networks assumes no responsibility for any inaccuracies in this document. Juniper Networks reserves the right to change, modify, transfer, or otherwise revise this publication without notice. How to Configure the NFX250 Copyright © 2021 Juniper Networks, Inc. All rights reserved. The information in this document is current as of the date on the title page. YEAR 2000 NOTICE Juniper Networks hardware and software products are Year 2000 compliant. Junos OS has no known time-related limitations through the year 2038. However, the NTP application is known to have some difficulty in the year 2036. END USER LICENSE AGREEMENT The Juniper Networks product that is the subject of this technical documentation consists of (or is intended for use with) Juniper Networks software. Use of such software is subject to the terms and conditions of the End User License Agreement ("EULA") posted at https://support.juniper.net/support/eula/. By downloading, installing or using such software, you agree to the terms and conditions of that EULA. iii Table of Contents About This Guide | ix 1 Overview NFX250 Overview | 2 JDM
    [Show full text]
  • Administering IBM Platform LSF Chapter 1
    Platform LSF Version 9 Release 1.3 Administering Platform LSF SC27-5302-03 Platform LSF Version 9 Release 1.3 Administering Platform LSF SC27-5302-03 Note Before using this information and the product it supports, read the information in “Notices” on page 827. First edition This edition applies to version 9, release 1 of IBM Platform LSF (product number 5725G82) and to all subsequent releases and modifications until otherwise indicated in new editions. Significant changes or additions to the text and illustrations are indicated by a vertical line (|) to the left of the change. If you find an error in any Platform Computing documentation, or you have a suggestion for improving it, please let us know. In the IBM Knowledge Center, add your comments and feedback to any topic. You can also send your suggestions, comments and questions to the following email address: [email protected] Be sure include the publication title and order number, and, if applicable, the specific location of the information about which you have comments (for example, a page number or a browser URL). When you send information to IBM, you grant IBM a nonexclusive right to use or distribute the information in any way it believes appropriate without incurring any obligation to you. © Copyright IBM Corporation 1992, 2014. US Government Users Restricted Rights – Use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM Corp. Contents Chapter 1. Managing Your Cluster . 1 Job Directories and Data..........441 Working with Your Cluster .........1 Resource
    [Show full text]
  • Experimental Study of Remote Job Submission and Execution on LRM Through Grid Computing Mechanisms
    Experimental Study of Remote Job Submission and Execution on LRM through Grid Computing Mechanisms Harshadkumar B. Prajapati Vipul A. Shah Information Technology Department Instrumentation & Control Engineering Dharmsinh Desai University Department Nadiad, INDIA Dharmsinh Desai University e-mail: [email protected], Nadiad, INDIA [email protected] e-mail: [email protected], [email protected] Abstract —Remote job submission and execution is The fundamental unit of work-done in Grid fundamental requirement of distributed computing done computing is successful execution of a job. A job in using Cluster computing. However, Cluster computing Grid is generally a batch-job, which does not interact limits usage within a single organization. Grid computing with user while it is running. Higher-level Grid environment can allow use of resources for remote job services and applications use job submission and execution that are available in other organizations. This execution facilities that are supported by a Grid paper discusses concepts of batch-job execution using LRM and using Grid. The paper discusses two ways of middleware. Therefore, it is very important to preparing test Grid computing environment that we use understand how a job is submitted, executed, and for experimental testing of concepts. This paper presents monitored in Grid computing environment. experimental testing of remote job submission and Researchers and scientists who are working in higher- execution mechanisms through LRM specific way and level Grid services and applications do not pay much Grid computing ways. Moreover, the paper also discusses attention to the underlying Grid infrastructure in their various problems faced while working with Grid published work.
    [Show full text]
  • The X-Bone & Its Virtual Internet Architecture
    The X-Bone & its Virtual Internet Architecture 10 Years Later Joe Touch, Greg Finn, Lars Eggert, Amy Hughes and Yu-Shun Wang Workshop on Overlay and Network Virtualization 16th GI/ITG Conference on Kommunikation in Verteilten Systemen Kassel, Germany March 6, 2009 March 6, 2009 Copyright 2009, USC/ISI. All rights reserved. 1 History X-Bone was a series of research projects at USC/ISI 1997-2005+ initial funding from DARPA, follow-on funding from the NSF http://www.isi.edu/xbone/ key results an architecture (the “Virtual Internet” architecture) a deployment/management system (the “X-Bone”) follow-on work using virtual nets March 6, 2009 Copyright 2009, USC/ISI. All rights reserved. 2 X-Bone Overlay System Web GUI Multiple views IP Base Star Overlay B A C D ring-ovl star-ovl B B A A Ring Overlay C D C D xd GUI Overlay Manager Resource Base IPv4 Resource Daemon Daemon Network Resource Daemon link host router X-Bone system Automated monitoring March 6, 2009 Copyright 2009, USC/ISI. All rights reserved. 3 X-Bone Timeline 1997 – first whitepaper 2001-2003 – NetFS (NSF) 1998-2001 – X-Bone (DARPA) File system configuration of IP overlays with revisitation, network properties recursion (LISP) 2002-2005 – X-Tend (NSF) 2000 – running code (FreeBSD, X-Bone for testbed uses Linux) 2003-2005 – DataRouter (int.) 2000 – application deployment Support for overlay P2P 2001 – TetherNet “NAT-buster” forwarding to support demos 2005-2006 – Agile Tunnels (NSA) 2001-2004 – DynaBone Partial overlays for DDOS safety (DARPA) 2006-2009 – RNA (NSF) 800-way spread-spectrum Extending X-Bone Choices model parallel overlays to general protocol stack 15-level deep overlays architecture March 6, 2009 Copyright 2009, USC/ISI.
    [Show full text]
  • Turbo-Charging Open Source Hadoop for Faster, More Meaningful Insights
    Turbo-Charging Open Source Hadoop for Faster, more Meaningful Insights Gord Sissons Senior Manager, Technical Marketing IBM Platform Computing [email protected] Agenda • Some Context – IBM Platform Computing • Low-latency scheduling meets open-source • Breakthrough performance • Multi-tenancy (for real!) • Cluster-sprawl - The elephant in the room • Side step the looming challenges IBM Platform Computing De facto standard for commercial high- • Acquired by IBM in 2012 performance computing Powers financial • 20 year history in high-performance computing analytics grids for 60% of top investment banks • 2000+ global customers Over 5 million CPUs • 23 of 30 largest enterprises under management • High-performance, mission-critical, extreme scale Breakthrough performance in Big • Comprehensive capability Data analytics Technical Computing - HPC Platform LSF Scalable, comprehensive workload Family management for demanding heterogeneous environments Simplified, integrated HPC management Platform HPC software bundled with systems Analytics Infrastructure Software High-throughput, low-latency compute Platform and data intensive analytics Symphony Family • An SOA infrastructure for analytics • Extreme performance and scale • Complex Computations (i.e., risk) • Big Data Analytics via MapReduce Our worldview – shaped by time critical analytics • Financial firms compete on the their ability to maximize use of capital • Monte-Carlo simulation is a staple technique for simulating market outcomes • Underlying instruments are increasingly complex •
    [Show full text]
  • Multihoming with ILNP in Freebsd
    MULTIHOMING WITH ILNP IN FREEBSD Bruce Simpson A Thesis Submitted for the Degree of PhD at the University of St Andrews 2016 Full metadata for this item is available in St Andrews Research Repository at: http://research-repository.st-andrews.ac.uk/ Please use this identifier to cite or link to this item: http://hdl.handle.net/10023/8681 This item is protected by original copyright This item is licensed under a Creative Commons Licence Multihoming with ILNP in FreeBSD Thesis by Bruce Simpson In Partial Fulfillment of the Requirements for the Degree of Doctor of Philosophy University of St Andrews School of Computer Science 2016 (Defended 17th December 2015) Declaration I, Bruce Simpson, certify that this thesis, which is approximately 40,000 words in length, has been written by me, that it is the record of work carried out by me, or principally by myself in collaboration with others as acknowledged, and that it has not been submitted in any previous application for a higher degree. I was admitted as a research student in October, 2011 and as a candidate for the degree of Doctor of Philosophy in November, 2014; the higher study for which this is a record was carried out in the University of St Andrews between 2011 and 2015. Date ................ Signature of candidate ............................... I hereby certify that the candidate has fulfilled the conditions of the Resolution and Regulations appropriate for the degree of Doctor of Philosophy in the University of St Andrews and that the candidate is qualified to submit this thesis in application for that degree.
    [Show full text]
  • Platform LSF Command Reference Chapter 99
    Platform LSF Version 9 Release 1.3 Command Reference SC27-5305-03 Platform LSF Version 9 Release 1.3 Command Reference SC27-5305-03 Note Before using this information and the product it supports, read the information in “Notices” on page 521. First edition This edition applies to version 9, release 1 of IBM Platform LSF (product numbers 5725G82 and 5725L25) and to all subsequent releases and modifications until otherwise indicated in new editions. Significant changes or additions to the text and illustrations are indicated by a vertical line (|) to the left of the change. If you find an error in any Platform Computing documentation, or you have a suggestion for improving it, please let us know. In the IBM Knowledge Center, add your comments and feedback to any topic. You can also send your suggestions, comments and questions to the following email address: [email protected] Be sure include the publication title and order number, and, if applicable, the specific location of the information about which you have comments (for example, a page number or a browser URL). When you send information to IBM, you grant IBM a nonexclusive right to use or distribute the information in any way it believes appropriate without incurring any obligation to you. © Copyright IBM Corporation 1992, 2014. US Government Users Restricted Rights – Use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM Corp. Contents Chapter 1. bacct ...........1 Chapter 23. blcollect ........165 Chapter 2. badmin ..........15 Chapter 24. blcstat .........167 Chapter 3. bapp ...........31 Chapter 25. blhosts .........169 Chapter 4.
    [Show full text]
  • IP Flow Mobility Support for Proxy Mobile Ipv6 Based Networks
    UNIVERSIT A` DEGLI STUDI DI PADOVA Facolt`adi Ingegneria Corso di Laurea in Ingegneria delle Telecomunicazioni Tesi di Laurea IP Flow Mobility support for Proxy Mobile IPv6 based networks Relatori: Prof. Michele Zorzi Dott. Telemaco Melia Candidato: Fabio Giust Anno accademico 2010-2011 Summary The ability of offloading selected IP data traffic from 3G to WLAN access networks is considered a key feature in the upcoming 3GPP specifications, being the main goal to alleviate data congestion in cellular networks while delivering a positive user experience. Lately, the 3GPP has adopted solutions that enable mobility of IP-based wireless devices relocating mobility functions from the terminal to the network. To this end, the IETF has standardized Proxy Mobile IPv6 (PMIPv6), a protocol capable to hide often complex mobility procedures from the mobile devices. This thesis, in line with the mentioned offload requirement, further extends PMI- Pv6 to support dynamic IP flow mobility management across access wireless net- works according to operator policies. In this work, we assess the feasibility of the proposed solution and provide an experimental analysis based on a prototype net- work setup, implementing the PMIPv6 protocol and the related enhancements for flow mobility support. La capacit`adi spostare flussi IP da una rete di accesso 3G ad una di tipo WLAN `econsiderata una caratteristica chiave nelle specifiche future di 3GPP, essendo il principale metodo per alleviare la congestione nelle reti cellulari mantenendo al contempo una ragionevole qualit`apercepita dall’utente. Recentemente, 3GPP ha adottato soluzioni di mobilit`aper dispositivi con accesso radio basato su IP, tra- slando le funzioni di supporto dal terminale alla rete, e, a questo scopo, IETF ha standardizzato Proxy Mobile IPv6 (PMIPv6), un protocollo studiato per nascondere le procedure di mobilit`aai sistemi mobili.
    [Show full text]