Openstack Installation

Total Page:16

File Type:pdf, Size:1020Kb

Openstack Installation TECH GUIDE OpenStack Installation: A Practitioner’s Guide Introduction TECH GUIDE: OpenStack Installation: A Practitioner’s Guide One of the great things about OpenStack is all the options you have for deploying it – from homebrew to hosted OpenStack to vendor appliances to OpenStack-as-a-service. Previously, Platform9 published a tech guide comparing various OpenStack deployment models. If you opt for a do- it-yourself (DIY) approach, then you face the question of which tool to use. This guide will familiarize you with the landscape of OpenStack installation tools, including an overview of the most popular ones: DevStack, RDO Packstack, OpenStack-Ansible, Fuel and TripleO. OpenStack Architecture Overview If you’re new to OpenStack it may be helpful to review the OpenStack components. (Skip this section if you’re already familiar with OpenStack.) OpenStack’s design, inspired by Amazon Web Services (AWS), has well-documented REST APIs that enable a self-service, elastic Infrastructure-as-a Service (IaaS) cloud. In addition, OpenStack is fundamentally agnostic to the underlying infrastructure and integrates well with various compute, virtualization, network and storage technologies. Heat Clarity UI CLI / Tools Scripts (orchestration) Nova Compute Scheduler (images) Cinder Keystone Glance (identity) Basic Storage Basic Network Neutron Network Block Controller Storage Figure 1: OpenStack architecture is loosely coupled, and extensible to support any hypervisor / container, storage, and network system 2 Plan to get NO help to install & manage YES OpenStack? Want to use Have budget NO existing YES NO for Professional YES infrastructure? Services? Use OpenStack- as-a-service Platform9 Need high Prefer to NO customization? YES NO own infrastructure? YES DIY OpenStack install Use hosted DevStack OpenStack RDO Packstack IBM BlueBox OpenStack-Ansible Rackspace Fuel TripleO Prefer minimal OpenStack Browneld NO lifecycle management? YES NO Environment? YES Use OpenStack Use OpenStack- Use OpenStack Use OpenStack- vendor distro as-a-service vendor distro as-a-service Red Hat Platform9 Red Hat Platform9 Rackspace Rackspace Mirantis Mirantis Short NO timeline to build private YES cloud? DIY OpenStack install DevStack Avoid RDO Packstack NO Vendor lock-in? YES OpenStack-Ansible Fuel TripleO Use OpenStack Use OpenStack- appliances as-a-service Cisco Metapod Platform9 Stratoscale ZeroStack VMware VIO Figure 2: How to choose an OpenStack Deployment Model 3 How to Choose an OpenStack Deployment Model The primary question that drives the choice of deployment models is whether your IT team has the expertise, and the inclination, to install and manage OpenStack. Depending on the desire to host your infrastructure and avoid vendor lock-in, various deployment models are available. If there’s a need for a high degree of customization, along with the flexibility to choose hypervisors, then a DIY install is probably the best option (see highlighted section of the flowchart on previous page). If you choose DIY install, there’s a wide choice of open source tools that are very easy to use and can create environments for use in development, testing or production. These tools can deploy OpenStack on bare metal, virtual machines or even containers. Some even install OpenStack in a production-grade, highly available architecture. But which tools are best suited for your requirements? Read on for an overview of some of the most popular tools, followed by a handy comparison matrix to summarize the options. More detailed documentation is available on each tool’s dedicated website.¹ OpenStack Installation: DevStack DevStack is a series of extensible scripts used to quickly bring up a complete OpenStack environment suitable for non-production use. It’s used interactively as a development environment. Since DevStack installs all-in-one OpenStack environments, it can be used to deploy OpenStack on a single VM, a physical server or a single LXC container. Each option is suitable depending on the hardware capacity available and the degree of isolation required. A multi-node OpenStack environment can also be deployed using DevStack, but that’s not a thoroughly tested use case. For either kind of setup, the steps involve installing a minimal version of one of the supported Linux distributions and downloading the DevStack Git repository. The repo contains a script stack.sh that must be run as a non-root user and will perform the complete install based on configuration settings. The official approved and tested Linux distributions are Ubuntu (LTS plus current dev release), Fedora (latest and previous release) and CentOS/RHEL 7 (latest major release). The supported databases are MySQL and PostgreSQL. RabbitMQ and Qpid are the recom- mended messaging service along with Apache as the web server. The setup defaults to a FlatDHCP network using Nova Network or a similar configuration in Neutron. The default services configured by DevStack are Keystone, Swift, Glance, Cinder, Nova, Nova Networking, Horizon and Heat. DevStack supports a plugin architecture to include additional services that are not included directly in the install. Summary of the Installation Process 1. Install one of the supported Linux Distributions 2. Download DevStack from git 3. git clone https://git.openstack.org/openstack-dev/devstack 4. Make any desired changes to the configuration 5. Add a non-root user, with sudo enabled, to run the install script 6. devstack/tools/create-stack-user.sh; su stack 7. Run the install and go grab a coffee 8. cd devstack /stack.sh 4 Configuration Options DevStack provides a bunch of configuration options that can be modified as needed. The sections below summarize some of the important ones. local.conf DevStack configuration is modified via the file local.conf. It’s a modified .ini format file that introduces a meta-section header to carry additional information regarding the configuration files to be changed. The new header is similar to [[‘ <phase> ‘|’ <config–file–name> ‘]]’, where <phase> is one of a set of phase names defined by stack. sh and <config-file-name> is the configuration filename. If the path of the config file does not exist, it is skipped. The file is processed strictly in sequence and any repeated settings will override previous values. The defined phases are: • local – extracts localrc from local.conf before stackrc is sourced • post-config – runs after the layer 2 services are configured and before they are started • extra – runs after services are started and before any files in extra.d are executed • post-extra – runs after files in extra.d are executed [[post-config|$NOVA_CONF]] [DEFAULT] use_syslog = True [osapi_v3] enabled = False A specific meta-section local|localrc is used to provide a default localrc file. This allows all custom settings for DevStack to be con- tained in a single file. If localrc exists it will be used instead to preserve backward compatibility. [[local|localrc]] FIXED_RANGE=10.20.30.40/49 ADMIN_PASSWORD=secret LOGFILE=$DEST/logs/stack.sh.log 5 openrc openrc configures login credentials suitable for use with the OpenStack command-line tools. openrc sources stackrc at the beginning in order to pick up HOST_IP and/or SERVICE_HOST to use in the endpoints. The values shown below are the default values. OS_PROJECT_NAME=demo OS_USERNAME=demo OS_PASSWORD=secret #The usual cautions about putting passwords in environment variables apply HOST_IP=127.0.0.1 #Typically set in thelocalrc section SERVICE_HOST=$HOST_IP OS_AUTH_URL=http://$SERVICE_HOST:5000/v2.0 #commented out by default # export KEYSTONECLIENT_DEBUG=1 # export NOVACLIENT_DEBUG=1 Minimal Configuration While stack.sh can run without a localrc section in local.conf, it’s easier to repeat installs by setting a few minimal variables. Below is an example of a minimal configuration for values that are often modified. Note: if the *_PASSWORD variables are not set, the install script will prompt for values: • No logging • Pre-set the passwords to prevent interactive prompts • Move network ranges away from the local network • Set the host IP if detection is unreliable [[local|localrc]] ADMIN_PASSWORD=secret DATABASE_PASSWORD=$ADMIN_PASSWORD RABBIT_PASSWORD=$ADMIN_PASSWORD SERVICE_PASSWORD=$ADMIN_PASSWORD #FIXED_RANGE=172.31.1.0/24 #FLOATING_RANGE=192.168.20.0/25 #HOST_IP=10.3.4.5 # export NOVACLIENT_DEBUG=1 Service Repositories The Git repositories used to check out the source for each service are controlled by a pair of variables set for each service. *_REPO points to the repository and *_BRANCH selects which branch to check out. These may be overridden in local.conf to pull source from a different repo. GIT_BASE points to the primary repository server. NOVA_REPO=$GIT_BASE/openstack/nova.git NOVA_BRANCH=master 6 The installation directory defaults to /opt/stack and can be set by the DEST parameter. DEST=/opt/stack Logging By default stack.sh output is only written to the console where it runs. It can be sent to a file in addition to the console by setting LOGFILE to the fully qualified name of the destination log file. Old log files are cleaned automatically if LOGDAYS is set to the number of days to keep old log files. DevStack will log the stdout output of the services it starts. When using screen this logs the output in the screen windows to a file. Without screen this simply redirects stdout of the service process to a file inLOGDIR . Some of the project logs will be colorized by default and can be turned off as below. LOGFILE=$DEST/logs/stack.sh.log LOGDAYS=1 LOG_COLOR=False LOGDIR=$DEST/logs Logging all services to a single syslog can be convenient. If the destination log host is not localhost, the settings below can be used to direct the message stream to the log host. SYSLOG=True SYSLOG_HOST=$HOST_IP SYSLOG_PORT=516 Database Backend The available databases are defined in the lib/databases directory. MySQL is the default database but can be replaced in the localrc section: disable_service mysql enable_service postgresql Messaging Backend Support for RabbitMQ is included. Additional messaging backends may be available via external plugins.
Recommended publications
  • Storm Clouds Platform: a Cloud Computing Platform for Smart City Applications
    RESEARCH ARTICLE Storm Clouds Platform: a cloud computing platform for smart city applications Marco Battarra, Marco Consonni*, Samuele De Domenico and Andrea Milani Hewlett Packard Italiana S.R.L., 9, Via Di Vittorio Giuseppe – 20063 Cernusco Sul Naviglio, Italy Abstract: This paper describes our work on STORM CLOUDS[1], a project with the main objective of migrating smart-city services, that Public Authorities (PAs) currently provided using traditional Information Technology, to a cloud-based environment. Our organization was in charge of finding the technical solutions, so we designed and im- plemented a cloud computing solution called Storm Clouds Platform (SCP), for that purpose. In principle, the applica- tions we ported could run on a public-cloud service, like Amazon Web ServicesTM[2] or Microsoft® Azure[3], that pro- vide computational resources on a pay-per-use paradigm. However, these solutions have disadvantages due to their proprietary nature: vendor lock-in is one of the issues but other serious problems are related to the lack of full control on how data and applications are processed in the cloud. As an example, when using a public cloud, the users of the cloud services have very little control on the location where applications run and data are stored, if there is any. This is identi- fied as one of the most important obstacles in cloud computing adoption, particularly in applications manage personal data and the application provider has legal obligation of preserving end user privacy[4]. This paper explains how we faced the problem and the solutions we found. We designed a cloud computing platform — completely based on open-software components — that can be used for either implementing private clouds or for porting applications to public clouds.
    [Show full text]
  • AIT Haproxy.Key
    HEIG-VD | TIC – Technologies de l’Information et de la Communication HAProxy (C) 2015 Marcel Graf HEIG-VD | TIC – Technologies de l’Information et de la Communication HAProxy ■ HAProxy (High Availability Proxy) is a load balancer implemented in software ■ Available as Open Source (GPL/LGPL license) (http://www.haproxy.org) ■ Available as commercial product (http://www.haproxy.com) ■ Also available as appliance: ALOHA ■ Runs on FreeBSD, Linux, OpenBSD and Solaris ■ Written by Willy Tarreau in 2000 ■ Willy is the maintainer of the Linux 2.4 kernel ■ Lives in Fontenay aux Roses ■ Used by high-profile websites: GitHub, Bitbucket, Stack Overflow, Reddit, Tumblr, Twitter 2 Administration IT | HAProxy | Année 2015/16 (C) 2015 Marcel Graf HEIG-VD | TIC – Technologies de l’Information et de la Communication HAProxy Features ■ HAProxy can be used for ■ Load balancing on TCP layer and HTTP layer ■ Normalization / filtering of TCP and HTTP traffic ■ HTTP rewriting ■ SSL offloading ■ HTTP compression offloading ■ Traffic regulation ■ Protection against DDoS and service abuse ■ … 3 Administration IT | HAProxy | Année 2015/16 (C) 2015 Marcel Graf HEIG-VD | TIC – Technologies de l’Information et de la Communication HAProxy operations ■ Health checks ■ HAProxy periodically sends probes to servers to check if they are still operational. A probe can be superficial or go deeper: ■ ping to server’s IP address (TCP mode) ■ TCP connection to server’s HTTP port (TCP mode) ■ HTTP request to server (HTTP mode) ■ Based on health checks HAProxy sets a server’s state to UP or DOWN ■ Server administrative state ■ The administrator can set a server into one of three administrative states ■ READY — Server is in normal mode, accepting requests ■ DRAIN — Removes server from load balancing, but still allows it to be health-checked and accept new persistent connections.
    [Show full text]
  • Cloud Onload Haproxy Cookbook
    Cloud Onload® HAProxy Cookbook The information disclosed to you hereunder (the “Materials”) is provided solely for the selection and use of Xilinx products. To the maximum extent permitted by applicable law: (1) Materials are made available "AS IS" and with all faults, Xilinx hereby DISCLAIMS ALL WARRANTIES AND CONDITIONS, EXPRESS, IMPLIED, OR STATUTORY, INCLUDING BUT NOT LIMITED TO WARRANTIES OF MERCHANTABILITY, NON-INFRINGEMENT, OR FITNESS FOR ANY PARTICULAR PURPOSE; and (2) Xilinx shall not be liable (whether in contract or tort, including negligence, or under any other theory of liability) for any loss or damage of any kind or nature related to, arising under, or in connection with, the Materials (including your use of the Materials), including for any direct, indirect, special, incidental, or consequential loss or damage (including loss of data, profits, goodwill, or any type of loss or damage suffered as a result of any action brought by a third party) even if such damage or loss was reasonably foreseeable or Xilinx had been advised of the possibility of the same. Xilinx assumes no obligation to correct any errors contained in the Materials or to notify you of updates to the Materials or to product specifications. You may not reproduce, modify, distribute, or publicly display the Materials without prior written consent. Certain products are subject to the terms and conditions of Xilinx’s limited warranty, please refer to Xilinx’s Terms of Sale which can be viewed at https://www.xilinx.com/legal.htm#tos; IP cores may be subject to warranty and support terms contained in a license issued to you by Xilinx.
    [Show full text]
  • Updates from the Open Quantum Safe Project
    Updates from the Open Quantum Safe Project Open Quantum Safe core team: Michael Baentsch Vlad Gheorghiu, evolutionQ & University of Waterloo Basil Hess, IBM Research Christian Paquin, Microsoft Research John Schanck, University of Waterloo Douglas Stebila, University of Waterloo Goutam Tamvada, University of Waterloo April 23, 2021 Abstract The Open Quantum Safe (OQS) project is an open-source project that aims to support the development and prototyping of quantum-resistant cryptography. This short note provides an update on the tools OQS makes available. 1 Introduction The Open Quantum Safe (OQS) project1 is an open-source project that aims to support the develop- ment and prototyping (of applications) of quantum-resistant cryptography. OQS consists of the following main lines of work: liboqs, an open source C library for quantum- resistant cryptographic algorithms, and prototype integrations into protocols and applications, including a fork of the widely used OpenSSL library. These tools support research by ourselves and others. To reduce the hurdle for getting started and to aid the uptake and use of these components, our tools are also available as ready-to-use binaries in the form of Docker images and test servers. In this short note, we provide an update on the Open Quantum Safe project. 2 liboqs liboqs is an open source C library for quantum-safe cryptographic algorithms. liboqs makes accessible a collection of open-source implementations of quantum-safe key encapsulation mechanism (KEM) and digital signature algorithms through a common API. liboqs builds on Linux, macOS, and Windows, on Intel, AMD, and ARM platforms. Some of the implementations of these algorithms have been directly contributed to liboqs by members of the NIST submission teams; others are incorporated from the PQClean project.
    [Show full text]
  • Red Hat Codeready Containers 1.19 Getting Started Guide
    Red Hat CodeReady Containers 1.19 Getting Started Guide Quick-start guide to using and developing with CodeReady Containers Last Updated: 2020-12-04 Red Hat CodeReady Containers 1.19 Getting Started Guide Quick-start guide to using and developing with CodeReady Containers Kevin Owen [email protected] Legal Notice Copyright © 2020 Red Hat, Inc. The text of and illustrations in this document are licensed by Red Hat under a Creative Commons Attribution–Share Alike 3.0 Unported license ("CC-BY-SA"). An explanation of CC-BY-SA is available at http://creativecommons.org/licenses/by-sa/3.0/ . In accordance with CC-BY-SA, if you distribute this document or an adaptation of it, you must provide the URL for the original version. Red Hat, as the licensor of this document, waives the right to enforce, and agrees not to assert, Section 4d of CC-BY-SA to the fullest extent permitted by applicable law. Red Hat, Red Hat Enterprise Linux, the Shadowman logo, the Red Hat logo, JBoss, OpenShift, Fedora, the Infinity logo, and RHCE are trademarks of Red Hat, Inc., registered in the United States and other countries. Linux ® is the registered trademark of Linus Torvalds in the United States and other countries. Java ® is a registered trademark of Oracle and/or its affiliates. XFS ® is a trademark of Silicon Graphics International Corp. or its subsidiaries in the United States and/or other countries. MySQL ® is a registered trademark of MySQL AB in the United States, the European Union and other countries. Node.js ® is an official trademark of Joyent.
    [Show full text]
  • Lighthouse Documentation Release 1.0.0
    lighthouse Documentation Release 1.0.0 William Glass March 07, 2016 Contents 1 Overview 3 2 Development 5 3 License 7 3.1 Getting Started..............................................7 3.2 Configuration...............................................8 3.3 Examples................................................. 17 3.4 Writing Plugins.............................................. 24 3.5 Source Docs............................................... 28 3.6 Release Notes.............................................. 41 Python Module Index 45 i ii lighthouse Documentation, Release 1.0.0 Lighthouse is a service node discovery system written in python, built with resilience, flexibility and ease- of-use in mind and inspired by Airbnb’s SmartStack solution. Out of the box it supports discovery via Zookeeper with cluster load balancing handled by an automatically configured HAProxy. To jump right in see the Getting Started page, or if you’d like to see it in action check out the Examples page. Contents 1 lighthouse Documentation, Release 1.0.0 2 Contents CHAPTER 1 Overview A lighthouse setup consists of three parts running locally on each node: a load balancer, the lighthouse-writer script and (usually) the lighthouse-reporter script. In a Lighthouse setup, no node’s application code is aware of the existence of other nodes, they talk to a local port handled by an instance of the load balancer which in turn routes traffic among the various known other nodes. This local load balancer is automatically updated when nodes come and go via the lighthouse-writer script, which talks to the discovery method (e.g. Zookeeper) to keep track of which nodes on which clusters are up. The lighthouse-reporter script likewise talks to the discovery method, it is responsible for running health checks on any services on the local node and reports to the discovery method that the healthy services are up and the unhealthy ones are down.
    [Show full text]
  • Oracle® Openstack Configuration Guide for Release 4.0
    Oracle® OpenStack Configuration Guide for Release 4.0 E90982-03 February 2018 Oracle Legal Notices Copyright © 2018, Oracle and/or its affiliates. All rights reserved. This software and related documentation are provided under a license agreement containing restrictions on use and disclosure and are protected by intellectual property laws. Except as expressly permitted in your license agreement or allowed by law, you may not use, copy, reproduce, translate, broadcast, modify, license, transmit, distribute, exhibit, perform, publish, or display any part, in any form, or by any means. Reverse engineering, disassembly, or decompilation of this software, unless required by law for interoperability, is prohibited. The information contained herein is subject to change without notice and is not warranted to be error-free. If you find any errors, please report them to us in writing. If this is software or related documentation that is delivered to the U.S. Government or anyone licensing it on behalf of the U.S. Government, then the following notice is applicable: U.S. GOVERNMENT END USERS: Oracle programs, including any operating system, integrated software, any programs installed on the hardware, and/or documentation, delivered to U.S. Government end users are "commercial computer software" pursuant to the applicable Federal Acquisition Regulation and agency-specific supplemental regulations. As such, use, duplication, disclosure, modification, and adaptation of the programs, including any operating system, integrated software, any programs installed on the hardware, and/or documentation, shall be subject to license terms and license restrictions applicable to the programs. No other rights are granted to the U.S. Government. This software or hardware is developed for general use in a variety of information management applications.
    [Show full text]
  • The Haproxy Guide to Multi-Layer Security Defense in Depth Using the Building Blocks of Haproxy
    The HAProxy Guide to Multi-Layer Security Defense in Depth Using the Building Blocks of HAProxy Chad Lavoie Table of Contents Our Approach to Multi-Layer Security 4 Introduction to HAProxy ACLs 6 Formatting an ACL 7 Fetches 11 Converters 12 Flags 13 Matching Methods 14 Things to do with ACLs 16 Selecting a Backend 18 Setting an HTTP Header 20 Changing the URL 21 Updating Map Files 21 Caching 23 Using ACLs to Block Requests 23 Updating ACL Lists 26 Conclusion 27 Introduction to HAProxy Stick Tables 28 Uses of Stick Tables 29 Defining a Stick Table 31 Making Decisions Based on Stick Tables 44 Other Considerations 49 Conclusion 54 Introduction to HAProxy Maps 55 The Map File 56 Modifying the Values 60 The HAProxy Guide to Multi-Layer Security 2 Putting It Into Practice 68 Conclusion 72 Application-Layer DDoS Attack Protection 73 HTTP Flood 74 Manning the Turrets 75 Setting Request Rate Limits 77 Slowloris Attacks 81 Blocking Requests by Static Characteristics 82 Protecting TCP (non-HTTP) Services 86 The Stick Table Aggregator 89 The reCAPTCHA and Antibot Modules 90 Conclusion 93 Bot Protection with HAProxy 94 HAProxy Load Balancer 95 Bot Protection Strategy 96 Beyond Scrapers 105 Whitelisting Good Bots 109 Identifying Bots By Their Location 111 Conclusion 114 The HAProxy Enterprise WAF 1​15 A Specific Countermeasure 1​16 Routine Scanning 1​17 HAProxy Enterprise WAF 1​24 Retesting with WAF Protection 1​26 Conclusion 1​29 The HAProxy Guide to Multi-Layer Security 3 Our Approach to Multi-Layer Security D​efending your infrastructure can involve a dizzying number of components: from network firewalls to intrusion-detection systems to access control safeguards.
    [Show full text]
  • Switching to Oracle Linux – a Total Value Analysis 1
    Technology Insight Paper Switching to Oracle Linux—a Total Value Analysis By John Webster, Sr. Analyst at Evaluator Group June 2020 Enabling you to make the best technology decisions Switching to Oracle Linux – a Total Value Analysis 1 © 2020 Evaluator Group, Inc. All rights reserved. Reproduction of this publication in any form without prior written permission is prohibited. Switching to Oracle Linux – a Total Value Analysis 2 Executive Summary Linux is now a foundational operating system for enterprise IT organizations. Linux allows them to deploy a secure, full-featured, cloud native-compatible and production class operating system that is based on open source development processes. In fact, according to the Linux Foundation, Linux is the primary platform for building the cloud and is experiencing continual year-over-year growth. One reason this is the case is that a majority of enterprise IT organizations feel that Linux is more secure than most other operating systems.1 Presently, enterprises have choices when it comes to acquiring Linux and related support services, the most common of these being Red Hat with Red Hat Enterprise Linux (RHEL). However, many enterprises, and especially Oracle database and application users, are now leveraging and in many cases consolidating, all of their Linux-based applications to Oracle Linux. This move simplifies their overall Linux application environment, increases security, while further reducing costs associated with a Linux operational and cloud-consistent environment. We have heard Oracle Linux users refer to this trend as a “One Linux” strategy. Here we review Oracle Linux which is free to download, use, and redistribute without a support contract.
    [Show full text]
  • Administration Guide Administration Guide SUSE Linux Enterprise High Availability Extension 12 SP4 by Tanja Roth and Thomas Schraitle
    SUSE Linux Enterprise High Availability Extension 12 SP4 Administration Guide Administration Guide SUSE Linux Enterprise High Availability Extension 12 SP4 by Tanja Roth and Thomas Schraitle This guide is intended for administrators who need to set up, congure, and maintain clusters with SUSE® Linux Enterprise High Availability Extension. For quick and ecient conguration and administration, the High Availability Extension includes both a graphical user interface (GUI) and a command line interface (CLI). For performing key tasks, both approaches (GUI and CLI) are covered in detail in this guide. Thus, administrators can choose the appropriate tool that matches their needs. Publication Date: September 10, 2021 SUSE LLC 1800 South Novell Place Provo, UT 84606 USA https://documentation.suse.com Copyright © 2006–2021 SUSE LLC and contributors. All rights reserved. Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or (at your option) version 1.3; with the Invariant Section being this copyright notice and license. A copy of the license version 1.2 is included in the section entitled “GNU Free Documentation License”. For SUSE trademarks, see http://www.suse.com/company/legal/ . All other third-party trademarks are the property of their respective owners. Trademark symbols (®, ™ etc.) denote trademarks of SUSE and its aliates. Asterisks (*) denote third-party trademarks. All information found in this book has been compiled with utmost attention to detail. However, this does not guarantee complete accuracy. Neither SUSE LLC, its aliates, the authors nor the translators shall be held liable for possible errors or the consequences thereof.
    [Show full text]
  • Autonomic Fault Tolerance Using Haproxy in Cloud Environment
    Autonomic Fault Tolerance Using HAProxy in Cloud Environment Thesis submitted in partial fulfillment of the requirements for the award of degree of Master of Engineering in Software Engineering Submitted By Vishonika Kaushal (Roll No. 800931024) Under the supervision of: Mrs. Anju Bala (Assistant Professor) COMPUTER SCIENCE AND ENGINEERING DEPARTMENT THAPAR UNIVERSITY PATIALA – 147004 June 2011 Acknowledgement Many people have shared their time and expertise to help me accomplish my goal. First, I would like to sincerely thank my guide, Mrs. Anju Bala, Assistant Professor, Thapar University, Patiala for her constant support and guidance. Her instant responses to my countless inquiries have been invaluable and motivational. It was a great opportunity to work under her supervision. Many thanks to Dr. Inderveer Chana, Assistant Professor, Thapar University for her moral support and research environment she had facilitated for this work. Finally I wish to thank my family and friends for their immense love and encouragement throughout my life without which it would not have been possible to complete this work. Last but not the least I would like to thank God who have always been with me in my good and bad times. Vishonika Kaushal (800931024) ii Abstract Cloud computing, with its great potentials in low cost and on-demand services, is a promising computing platform for both commercial and non-commercial computation clients. In this thesis various fault tolerance techniques have been discussed to improve performance and availability of server applications. Cloud computing is a relatively new way of referring to the use of shared computing resources, and it is an alternative to having local servers handle applications.
    [Show full text]
  • ECS with Haproxy Load Balancer Deployment
    Technical White Paper ECS with HAProxy Load Balancer Abstract This document provides a technical deployment reference of the Dell EMC™ ECS™ with HAProxy load balancer. April 2021 H15785.4 Revisions Revisions Date Description January 2017 Initial release November 2017 Modified based on input from HAProxy April 2021 Update for HAProxy 2.1 with http head in Chapter 5.3.1 Acknowledgments Author: Zhu, Jarvis Support: Kraft, Chris The information in this publication is provided “as is.” Dell Inc. makes no representations or warranties of any kind with respect to the information in this publication, and specifically disclaims implied warranties of merchantability or fitness for a particular purpose. Use, copying, and distribution of any software described in this publication requires an applicable software license. Copyright ©2017-2021. Dell Inc. or its subsidiaries. All Rights Reserved. Dell, EMC, and other trademarks are trademarks of Dell Inc. or its subsidiaries. Other trademarks may be the property of their respective owners. Published in the USA [4/28/2021] [Document Type] [H15785.4] Dell believes the information in this document is accurate as of its publication date. The information is subject to change without notice. This document may contain language from third party content that is not under Dell's control and is not consistent with Dell's current guidelines for Dell's own content. When such third-party content is updated by the relevant third parties, this document will be revised accordingly. 2 ECS with HAProxy Load Balancer |
    [Show full text]