Monitoring Systems and POWER5/6 Lpars with Ganglia

Total Page:16

File Type:pdf, Size:1020Kb

Monitoring Systems and POWER5/6 Lpars with Ganglia Monitoring Systems and POWER5/6 LPARs with Ganglia Michael Perzl – [email protected] Agenda . Ganglia – what is it ? . Ganglia components and data flow . An introduction to RRDTool . Ganglia metrics – what can be measured ? . New POWER5/6 metrics (AIX & Linux) . Extending Ganglia with gmetric . Add device specific information to Ganglia . Ganglia network communication . Installation issues . Where to get Ganglia for AIX and Linux on POWER ? . Best practices . Future additions / plans . Discussion . Links 2 Monitoring Systems and POWER5/6 LPARs with Ganglia Ganglia – what is it ? Ganglia – what is it ? (1/3) . Ganglia is an Open Source cluster performance monitoring tool and has been extended to include POWER5/6 features like shared processor LPARs, entitlement, physical CPU usage etc. This session covers: – the technical details of Ganglia and the POWER5/6 extensions – how to set it up and use it to monitor all LPARs in a single machine and lots of machines 4 Monitoring Systems and POWER5/6 LPARs with Ganglia Ganglia – what is it ? (2/3) Ganglia properties: . scalable distributed monitoring system for high-performance computing systems such as clusters and grids . based on a hierarchical design targeted at federations of clusters . relies on a multicast-based listen/announce protocol to monitor state within clusters and uses a tree of point-to-point connections amongst representative cluster nodes to federate clusters and aggregate their state . leverages widely used technologies such as – XML for data representation – XDR (eXternal Data Representation) for compact, portable data transport – RRDtool for data storage and visualization . uses carefully engineered data structures and algorithms to achieve very low per-node overheads and high concurrency . robust implementation . Open Source, written in C – Downloaded 110,000+ times, 145+ countries, 500+ clusters, 2000+ nodes 5 Monitoring Systems and POWER5/6 LPARs with Ganglia Ganglia – what is it ? (3/3) Ganglia properties (cont.): . has been ported to an extensive set of operating systems and processor architectures: – AIX – Darwin – FreeBSD – HP-UX – IRIX – Linux – OSF – NetBSD – Solaris – Windows (via Cygwin) . is currently in use on over 500+ clusters around the world . has been used to link clusters across university campuses and around the world and can scale to handle clusters with 2000+ nodes – check http://ganglia.info/ for more details 6 Monitoring Systems and POWER5/6 LPARs with Ganglia Ganglia components and data flow Ganglia components The ganglia system consists of: . two unique daemons: – Ganglia Monitoring Daemon (gmond) • monitoring daemon, collects the metrics • runs on each node – Ganglia Meta Daemon (gmetad) • polls all gmond clients and stores the collected metrics in Round-Robin Databases (RRDs) . a PHP-based web frontend . a few other small utility programs – gmetric • can be used to easily extend Ganglia with additional user-defined metrics – gstat – gexec 8 Monitoring Systems and POWER5/6 LPARs with Ganglia Ganglia – Schematic View From: “Ganglia: Past, Present and Future” by Matt Massie: URL: http://ganglia.info/talks/lug_lbl_talk/ 9 Monitoring Systems and POWER5/6 LPARs with Ganglia Ganglia Architecture 10 Monitoring Systems and POWER5/6 LPARs with Ganglia Ganglia Monitoring Daemon (gmond) . Ganglia Monitoring Daemon (gmond) is a multi-threaded daemon which runs on each cluster node you want to monitor. Installation is easy: – just the daemon and a configuration file (/etc/gmond.conf) . gmond has four main responsibilities: 1. monitor changes in host state 2. announce relevant changes 3. listen to the state of all other ganglia nodes via a unicast or multicast channel 4. answer requests for an XML description of the cluster state . Each gmond transmits information in two different ways: – unicasting or multicasting host state in external data representation (XDR) format using UDP messages – sending XML over a TCP connection 11 Monitoring Systems and POWER5/6 LPARs with Ganglia Ganglia Meta Daemon (gmetad) (1/2) . Ganglia Meta Daemon (gmetad) is a daemon which typically only runs on one specific cluster node – or on more when using a staged setup. Installation is easy: – just the daemon and a configuration file (/etc/gmetad.conf) . Federation in Ganglia is achieved using a tree of point-to-point connections amongst representative cluster nodes to aggregate the state of multiple clusters. At each node in the tree a gmetad – periodically polls a collection of child data sources – parses the collected XML – saves all numeric volatile metrics to round-robin databases – exports the aggregated XML over a TCP socket to clients 12 Monitoring Systems and POWER5/6 LPARs with Ganglia Ganglia Meta Daemon (gmetad) (2/2) . Data sources may be either – gmond daemons, representing specific clusters or – other gmetad daemons, representing sets of clusters . Data sources use source IP addresses for access control – Multiple IP addresses can be specified for failover – The capability is natural for aggregating data from clusters since each gmond daemon contains the entire state of its cluster 13 Monitoring Systems and POWER5/6 LPARs with Ganglia Ganglia PHP web frontend (1/2) Web frontend properties: . provides a view of the gathered information via real-time dynamic web pages . displays Ganglia data in a meaningful way for system administrators and users – For example, one can view the CPU utilization over the past hour, day, week, month, or year – The web frontend shows similar graphs for memory usage, disk usage, network statistics, number of running processes, and all other Ganglia metrics 14 Monitoring Systems and POWER5/6 LPARs with Ganglia Ganglia PHP web frontend (2/2) Web frontend properties (cont.): . depends on the existence of the gmetad which provides it with data from several Ganglia sources . opens the local port 8651 (by default) and expects to receive a Ganglia XML tree . the web pages themselves are highly dynamic; any change to the Ganglia data appears immediately on the site – This behavior leads to a very responsive site, but requires that the full XML tree be parsed on every page access – Therefore, the Ganglia web frontend should run on a fairly powerful, dedicated machine if it presents a large amount of data . is written in the PHP scripting language and uses graphs generated by gmetad to display history information . has been tested on many flavors of Unix (primarily Linux) with the Apache web server and the PHP 4.1 module 15 Monitoring Systems and POWER5/6 LPARs with Ganglia Ganglia - data flow (1/4) One daemon per node/LPAR gmond Operating System performance stats /etc/gmond.conf API File access Network Web 16 Monitoring Systems and POWER5/6 LPARs with Ganglia Ganglia - data flow (2/4) One daemon per node/LPAR Runs on web server gmond gmetad /etc/gmetad.conf rrdtool Operating System performance stats /etc/gmond.conf database API of statistics Browser File access Network Web 17 Monitoring Systems and POWER5/6 LPARs with Ganglia Ganglia - data flow (3/4) One daemon per node/LPAR Runs on web server gmond gmetad /etc/gmetad.conf rrdtool Operating System performance stats /etc/gmond.conf database API of statistics Ganglia FE scripts Browser File access Apache2 Network + PHP5 Web 18 Monitoring Systems and POWER5/6 LPARs with Ganglia Ganglia - data flow (4/4) User command One daemon per node/LPAR Runs on web server gmetric gmond gmetad /etc/gmetad.conf rrdtool Operating System performance stats /etc/gmond.conf database API of statistics Ganglia FE scripts Browser File access Apache2 Network + PHP5 Web 19 Monitoring Systems and POWER5/6 LPARs with Ganglia Ganglia - data flow again One daemon per node/LPAR Only one instance with the Web Server /etc/gmetad.conf /etc/gmond.conf gmond gmetad rrdtool /etc/gmond.conf gmond database of statistics /etc/gmond.conf gmond PHP scripts Browser Apache2 File access + PHP5 Network Web 20 Monitoring Systems and POWER5/6 LPARs with Ganglia An introduction to RRDTool RRDTool . Homepage: http://oss.oetiker.ch/rrdtool/ . RRD is the Acronym for Round-Robin Database. RRD is a system to store and display time-series data (i.e., network bandwidth, machine-room temperature, server load average). It stores the data in a very compact way that will not expand over time (fixed size of DB), and it presents useful graphs by processing the data to enforce a certain data density. It can be used either via simple wrapper scripts (from shell or Perl) or via frontends that poll network devices and put a friendly user interface on it. RRDTool is the industry standard tool to store and display time-series data! 22 Monitoring Systems and POWER5/6 LPARs with Ganglia RRDTool example graph Graph taken from http://oss.oetiker.ch/rrdtool/gallery/index.en.html Graph shows inbound and outbound call traffic going in and out of the switch via the 6 trunks connected to the Diamond exchange. Inbound traffic shown as positive and uses a lowest-free fill method. Outbound traffic shown as negative uses a distributed fill method. Tech details on RRDtrac. 23 Monitoring Systems and POWER5/6 LPARs with Ganglia RRDTool example # rrdtool create test.rrd \ --start 920804400 \ --step 300 \ DS:km:COUNTER:600:U:U \ RRA:AVERAGE:0.5:1:24 # rrdtool update test.rrd 920804700:12345 920805000:12357 920805300:12363 # rrdtool update test.rrd 920805600:12363 920805900:12363 920806200:12373 # rrdtool update test.rrd 920806500:12383 920806800:12393 920807100:12399 # rrdtool update test.rrd 920807400:12405 920807700:12411 920808000:12415 # rrdtool update test.rrd 920808300:12420 920808600:12422 920808900:12423 # rrdtool graph kilometer.png \ --start 920804400 \ --end 920808000 \ DEF:mykm=test.rrd:km:AVERAGE \ LINE2:mykm#FF0000 24 Monitoring Systems and POWER5/6 LPARs with Ganglia Ganglia metrics – what can be monitored ? Metrics Definition of a metric: . A metric is a certain observed property of the system. Number of metrics: . 34 standard metrics, i.e., available (i.e., defined) on all platforms . Additional platform dependent metrics available – Solaris • 8 additional metrics available – HP-UX • 4 additional metrics available – AIX • 18 additional new metrics available for POWER5/6 !!! • details later….
Recommended publications
  • Monitoring Cluster on Online Compiler with Ganglia
    Monitoring Cluster on Online Compiler with Ganglia Nuryani, Andria Arisal, Wiwin Suwarningsih, Taufiq Wirahman Research Center for Informatics, Indonesian Institute of Sciences Komp. LIPI, Gd. 20, Lt. 3, Sangkuriang, Bandung, 40135 Indonesia Email : {nuryani|andria.arisal|wiwin|taufiq}@informatika.lipi.go.id Abstract strength and weaknesses in certain part compared with the others. For the example, Ganglia is more concerned Ganglia is an open source monitoring system for high with metrics gathering and tracking them over time performance computing (HPC) that collect both a whole while Nagios has focused in an alerting mechanism, cluster and every nodes status and report to the user. Visuel has an additional function to visualize the We use Ganglia to monitor our performance data of MPI application and Clumon spasi.informatika.lipi.go.id (SPASI), a customized- offers web-based pages to show the detailed fedora10-based cluster, for our cluster online compiler, information about resources, queues, jobs and CLAW (cluster access through web). Our experience on warnings[2, 3, 6]. using Ganglia shows that Ganglia has a capability to At present, we are developing online compiler for view our cluster status and allow us to track them. high performance computing environment focusing on Keywords: high performance computing, cluster, cluster developing web-interface to ease user (especially monitoring, ganglia programmer) on using cluster infrastructure named . CLAW (cluster access through web). Online compiler 1. Introduction is a web-based application that provides service for The need for high performance computer which writing, compilation, and execution of source code closely related with supercomputer and massively through web.
    [Show full text]
  • Setup of a Ganglia Monitoring System for a Grid Computing Cluster
    Setup of a Ganglia Monitoring System for a Grid Computing Cluster Aufbau eines Ganglia-Überwachungssystems für ein Grid-Computing-Cluster von Oleg Ivasenko Bachelorarbeit in Physik vorgelegt der Fakultät für Mathematik, Informatik und Naturwissenschaften der RWTH Aachen im September, 2015 angefertigt im Physikalischen Institut IIIb bei Prof. Dr. Achim Stahl Zusammenfassung Das RWTH Grid Cluster ist ein Tier 2/3 Rechenzentrum im WLHC Grid mit über zwei hundert Rechnern. Diese müssen bei Betrieb überwacht werden, um aufkommende Probleme analysieren zu können. Da das aktuell laufende System gewisse Inkompatibilitäten zu neuen Versionen des laufenden Betriebssystems zeigt, soll es durch Ganglia ersetzt werden. In dieser Arbeit wird der Aufbau des Ganglia Monitoring Systems und die Anpassung dessen an die Besonderheiten des RWTH Grid Clusters vorgenommen. Das Ganglia Monitoring System überwacht das gesamte Cluster und zeigt sich weiterhin skalierbar für zukünftige Erweiterungen. Abstract The RWTH Grid Cluster is a Tier 2/3 grid site in the WLHC Grid with over two hundred com- puters. Those have to be monitored while operating to analyze occurring problems. The currently running monitoring system shows incompatibilities with newer versions of the running operating system and thus should be replaced by Ganglia. In this thesis the setup and the customization of the Ganglia Monitoring System to suit the RWTH Grid Cluster specic needs is detailed. The Ganglia Monitoring System monitors the whole Cluster and proves itself as further scalable for future extensions. Contents List of Figures I List of Tables II List of Listings II 1 Introduction 1 1.1 Worldwide LHC Computing Grid . .2 1.2 RWTH Grid Cluster .
    [Show full text]
  • Ganglia Users Guide 7.0 Edition Ganglia Users Guide 7.0 Edition Published Dec 01 2017 Copyright © 2017 University of California
    Ganglia Users Guide 7.0 Edition Ganglia Users Guide 7.0 Edition Published Dec 01 2017 Copyright © 2017 University of California This document is subject to the Rocks® License (see Appendix: Rocks Copyright). Table of Contents Preface.....................................................................................................................................................................v 1. Overview.............................................................................................................................................................1 2. Installing.............................................................................................................................................................2 2.1. On a New Server.....................................................................................................................................2 2.2. On an Existing Server..............................................................................................................................2 3. Using the Ganglia Roll.......................................................................................................................................4 3.1. Using the Ganglia Roll............................................................................................................................4 4. Customizing the Ganglia Roll...........................................................................................................................6 4.1. Customizing the Ganglia Web interface..................................................................................................6
    [Show full text]
  • Monitoring Systems and Tricks of the Trade
    Enabling Grids for E-sciencE Monitoring Systems and Tricks of the Trade Antun Balaz Scientific Computing Laboratory Institute of Physics Belgrade http://www.scl.rs/ 21 Jan – 01 Feb 2009 www.eu-egee.org FP7-INFRA-222667 The 2nd workshop on HPC, IPM and Shahid Beheshti University, Tehran, Iran Overview Enabling Grids for E-sciencE • Ganglia (fabric monitoring) • Nagios (fabric + network monitoring) • Yumit/Pakiti (security) • CGMT (integration + hardware sensors) • WMSMON (custom service monitoring) • BBmSAM (mobile interface) • CLI scripts • Summary FP7-INFRA-222667 The 2nd workshop on HPC, IPM and Shahid Beheshti University, Tehran, Iran Ganglia Overview Enabling Grids for E-sciencE • Introduction • Ganglia Architecture • Apache Web Frontend • Gmond & Gmetad • Extending Ganglia – GMetrics – Gmond Module Development FP7-INFRA-222667 The 2nd workshop on HPC, IPM and Shahid Beheshti University, Tehran, Iran Introduction to Ganglia Enabling Grids for E-sciencE • Scalable Distributed Monitoring System • Targeted at monitoring clusters and grids • Multicast-based Listen/Announce protocol • Depends on open standards – XML – XDR compact portable data transport – RRDTool - Round Robin Database – APR – Apache Portable Runtime – Apache HTTPD Server – PHP based web interface • http://ganglia.sourceforge.net or http://www.ganglia.info FP7-INFRA-222667 The 2nd workshop on HPC, IPM and Shahid Beheshti University, Tehran, Iran Ganglia Architecture Enabling Grids for E-sciencE • Gmond – Metric gathering agent installed on individual servers • Gmetad –
    [Show full text]
  • Ganglia Deployment Guide @CERN Crist´Ov˜Aocordeiro, IT-SDC-ID [email protected]
    Ganglia Deployment Guide @CERN Crist´ov~aoCordeiro, IT-SDC-ID [email protected] 1 Introduction With the constant increase of Cloud resources being used, there is the need of having something which allows their monitoring/accounting. This can be very useful and practical in a way where we will be able to not only create alarms, and automatic decisions based on these, but also compare and report resource usages for specific cloud infrastructures, over specific periods of time. Ganglia is a distributed system which allows us to achieve exactly this. "Ganglia is a scalable distributed monitoring system for high-performance computing systems such as clusters and Grids. It is based on a hierarchical design targeted at federations of clusters. It leverages widely used technolo- gies such as XML for data representation, XDR for compact, portable data transport, and RRDtool for data storage and visualization. It uses carefully engineered data structures and algorithms to achieve very low per-node over- heads and high concurrency. The implementation is robust, has been ported to an extensive set of operating systems and processor architectures, and is currently in use on thousands of clusters around the world."1. It works on a Headnode Node based interaction and its deployment is fairly easy. During instantiation the workernodes can also be easily config- ured to report their metrics to the headnode by using the Ganglia module for CloudInit which is available and maintained at CERN (configuration through other contextualization methods is also possible). NOTE: Ganglia is a highly configurable system. This guide only covers the basics for some specific situations.
    [Show full text]
  • The Ganglia Distributed Monitoring System: Design, Implementation, and Experience
    The Ganglia Distributed Monitoring System: Design, Implementation, and Experience Matthew L. Massie Brent N. Chun David E. Culler Univ. of California, Berkeley Intel Research Berkeley Univ. of California, Berkeley [email protected] [email protected] [email protected] Abstract growth, system configurations inevitably need to adapt. In summary, high performance systems today have sharply di- Ganglia is a scalable distributed monitoring system for high- verged from the monolithic machines of the past and now performance computing systems such as clusters and Grids. face the same set of challenges as that of large-scale dis- It is based on a hierarchical design targeted at federations of tributed systems. clusters. It relies on a multicast-based listen/announce proto- col to monitor state within clusters and uses a tree of point- One of the key challenges faced by high performance to-point connections amongst representative cluster nodes to distributed systems is scalable monitoring of system state. federate clusters and aggregate their state. It leverages widely Given a large enough collection of nodes and the associated used technologies such as XML for data representation, XDR computational, I/O, and network demands placed on them for compact, portable data transport, and RRDtool for data by applications, failures in large-scale systems become com- storage and visualization. It uses carefully engineered data monplace. To deal with node attrition and to maintain the structures and algorithms to achieve very low per-node over- health of the system, monitoring software must be able to heads and high concurrency. The implementation is robust, quickly identify failures so that they can be repaired either has been ported to an extensive set of operating systems and automatically or via out-of-band means (e.g., rebooting).
    [Show full text]
  • Thirty Billion Metrics a Day: Large-Scale Performance Metrics with Ganglia Adam Compton, Quantcast [email protected] @Comptona
    Thirty Billion Metrics a Day: Large-Scale Performance Metrics with Ganglia Adam Compton, Quantcast [email protected] @comptona November 8–13, 2015 | Washington, D.C. www.usenix.org/lisa15 #lisa15 Thirty Billion Metrics a Day: Forty Adam Compton, Quantcast [email protected] @comptona November 8–13, 2015 | Washington, D.C. www.usenix.org/lisa15 #lisa15 ● 7,000,000 unique metrics collected every 15 sec ● from 1000s of machines ● 480,000 metrics/second ● 41,000,000,000 (billion) metrics/day How many machines do we need for this? Two. Who am I? ● What is Ganglia? ● How Ganglia works ● Inputs ● Outputs ● Scaling ● What’s next What is Ganglia? “Ganglia is a scalable distributed monitoring system for high-performance computing systems” What does “monitoring” mean? What is Ganglia - monitoring vs. metrics performance metrics: regular, numeric, time-series data monitoring: scheduled checks (of metrics, processes, endpoints, etc.) to identify anomalous behavior alerting: getting a human’s attention visualization: charts and graphs By these definitions, Ganglia collects, aggregates, and visualizes performance metrics. What is Ganglia not? ● No monitoring ● No alerting ● No asset database ● No Julienne fries ● What is Ganglia? ● How Ganglia works ● Inputs ● Outputs ● Scaling ● What’s next How Ganglia works How Ganglia works - gmond How Ganglia works - gmetad (aside) How RRDs work How Ganglia works - gweb How Ganglia works - Upsides Built-in hierarchy, summarization, and metadata Web UI is easy to browse and grok How Ganglia works - Downsides
    [Show full text]
  • HPC Best Practices: Software Lifecycles San Francisco, California September 28-29, 2009
    Report of the 3rd DOE Workshop on HPC Best Practices: Software Lifecycles San Francisco, California September 28-29, 2009 HPC Best Practices: Software Lifecycles Findings and notes from the the third in a series of HPC community workshops Workshop Sponsors: DOE Office of Science – ASCR and DOE NNSA Chair: David E. Skinner (LBNL) Vice-Chair: Jon Stearley (SNL) Report Editors: John Hules (LBNL) and Jon Bashor (LBNL Attendees/Report Contributors: Deb Agarwal (LBNL), Dong Ahn (LLNL), William Allcock (ALCF), Kenneth Alvin (SNL), Chris Atwood (HPCMPO/CREATE), Charles Bacon (ANL), Bryan Biegel (NASA NAS), Buddy Bland (ORNL), Brett Bode (NCSA), Robert Bohn (NCO/NITRD), Jim Brandt (SNL), Jeff Broughton (NERSC), Shane Canon (NERSC), Shreyas Cholia (NERSC), Bob Ciotti (NASA NAS), Susan Coghlan (ANL), Bradley Comes (DoD HPCMP ), Paul “Steve” Cotter (ESnet), Nathan DeBardeleben (LANL), Sudip Dosanjh (SNL), Tony Drummond (LBNL), Mark Gary (LLNL), Ann Gentile (SNL), Gary Grider (LANL), Pam Hamilton (LLNL), Andrew Hanushevsky (SLAC), Daniel Hitchcock (DOE ASCR), Paul Iwanchuk (LANL), Gary Jung (LBNL), Suzanne Kelly (SNL), Patricia Kovatch (NICS/UTK), William Kramer (NCSA), Manojkumar Krishnan (PNNL), Sander Lee(NNSA), Janet Lebens (Cray Inc.), Lie-Quan Lee (SLAC), Ewing Lusk (ANL), Osni Marques (DOE ASCR), David Montoya (LANL), Krishna Muriki (LBNL), Tinu Ogunde (ASC/NNSA/DOE), Viraj Paropkari (NERSC), Rob Pennington (NSF), Terri Quinn (LLNL), Randal Rheinheimer (LANL), Alain Roy (Univ. Wisconsin-Madison), Stephen Scott (ORNL), Yukiko Sekine (DOE ASCR),
    [Show full text]
  • Ganglia on the Brain
    iVoyeur: More Ganglia on the Brain DAVE JOSEPHSEN Dave Josephsen is the Welcome to the second installment in my Ganglia series . In the last issue I prom- author of Building a ised (threatened?) to walk you through building a Ganglia data collection module Monitoring Infrastructure in C, and that’s exactly what we’re going to do . with Nagios (Prentice Ganglia, as you’ll recall, is composed primarily of two daemons: gmond, which Hall PTR, 2007) and is senior systems runs on the client side, collecting metrics, and gmetad, which runs server-side engineer at DBG, Inc., where he maintains and polls the clients, pushing their data into RRDTool databases . Gmond, out of a gaggle of geographically dispersed server the box, can collect a litany of interesting metrics from the system on which it is farms. He won LISA ’04’s Best Paper award for installed and can be extended to collect user-defined metrics in three different his co-authored work on spam mitigation, and ways . he donates his spare time to the SourceMage GNU Linux Project. The first method to add new metrics to gmond is to collect the data yourself using [email protected] a shell script, or similar program, piping the resultant data to the Gexec daemon . This method is simple and widely used, but adds overhead to a system carefully designed to minimize it . The second and third methods are to write plug-ins for gmond in Python or C . Writing gmond plug-ins in Python requires that Python be installed and that gmond be compiled against the Python libs on every box where you might want to use your module .
    [Show full text]
  • Open Source Software Options for Government
    Open Source Software Options for Government Version 2.0, April 2012 Aim 1. This document presents options for Open Source Software for use in Government. 2. It is presented in recognition that open source software is underused across Government and the wider public sector. 3. This set of options is primarily intended to be used by Government to encourage IT suppliers and integrators to evaluate open source options when designing solutions and services. 4. This publication does not imply preference for any vendor or product. Open source software, by definition, is not tied inextricably to any particular commercial organisation. Any commercial entity can choose to support, maintain, or integrate open source software. 5. It is understood that the software market, and the open source ecosystem in particular, is a rapidly developing environment and any options list will be incomplete and may become outdated quickly. Even so, given the relatively low level of open source experience in Government, this options list has proven useful for encouraging IT suppliers to consider open source, and to aid the assurance of their proposals. Context 1. The Coalition Government believes Open Source Software can potentially deliver significant short and long term cost savings across Government IT. 2. Typical benefits of open source software include lower procurement prices, no license costs, interoperability, easier integration and customisation, fewer barriers to reuse, conformance to open technology and data standards giving autonomy over your own information, and freedom from vendor lock in. 3. Open Source is not widely used in Government IT. The leading systems integrators and supplies to Government do not routinely and effectively consider open source software for IT solutions, as required by the existing HMG ICT policy.
    [Show full text]
  • GANGLIA on the BRAIN 55 Specific Purpose in Mind, So Some of What It Isn’T Is by Design
    IT’S IRONIC, I THINK, THAT I’VE BEEN dissatisfied with my RRDTool data DAVE JOSEPHSEN collection and display options for as long as I have. Ironic in the sense that RRDTool is such a category killer for what it does (storing and graphing time series data). iVoyeur: Ganglia It used to be novel, to see that distinctive- looking RRDTool graph popping up on the brain somewhere unexpected, but these days Dave Josephsen is the author of Building it’s so ubiquitous that I’m surprised to see a Monitoring Infrastructure with Nagios anything but. So if RRDTool is the ultimate (Prentice Hall PTR, 2007) and is senior systems engineer at DBG, Inc., where he solution for everything that happens maintains a gaggle of geographically dispersed server farms. He won LISA ’04’s between data coming in and graphs going Best Paper award for his co-authored out, the open source community has surely work on spam mitigation, and he do- nates his spare time to the SourceMage provided category killers for polling data GNU Linux Project. and displaying those graphs, yes? [email protected] No. When it comes to polling and displaying that data, we’re far from having a category-killer, in my opinion. It’s not for lack of candidates, mind you. Freshmeat can provide you with myriad options written in all manner of languages (but mostly PHP) that do essentially the same thing—put an RRDTool graph on a Web page. It’s a bit of a cop-out to say they all do “essentially” the same thing, in fact, since they all seem to make the same mistakes over and over again.
    [Show full text]
  • Maestro: a Remote Execution Tool for Visualization Clusters Aron Lee Bierbaum Iowa State University
    Iowa State University Capstones, Theses and Retrospective Theses and Dissertations Dissertations 2007 Maestro: a remote execution tool for visualization clusters Aron Lee Bierbaum Iowa State University Follow this and additional works at: https://lib.dr.iastate.edu/rtd Part of the Computer Sciences Commons Recommended Citation Bierbaum, Aron Lee, "Maestro: a remote execution tool for visualization clusters" (2007). Retrospective Theses and Dissertations. 14635. https://lib.dr.iastate.edu/rtd/14635 This Thesis is brought to you for free and open access by the Iowa State University Capstones, Theses and Dissertations at Iowa State University Digital Repository. It has been accepted for inclusion in Retrospective Theses and Dissertations by an authorized administrator of Iowa State University Digital Repository. For more information, please contact [email protected]. Maestro: A remote execution tool for visualization clusters by Aron Lee Bierbaum A thesis submitted to the graduate faculty in partial fulfillment of the requirements for the degree of MASTER OF SCIENCE Major: Computer Engineering Program of Study Committee: Carolina Cruz-Neira, Major Professor Julie Dickerson Chris Harding Iowa State University Ames, Iowa 2007 UMI Number: 1447485 UMI Microform 1447485 Copyright 2008 by ProQuest Information and Learning Company. All rights reserved. This microform edition is protected against unauthorized copying under Title 17, United States Code. ProQuest Information and Learning Company 300 North Zeeb Road P.O. Box 1346 Ann Arbor, MI 48106-1346
    [Show full text]