UNIT- I INTRODUCTION Grid Computing Values and Risks

Total Page:16

File Type:pdf, Size:1020Kb

UNIT- I INTRODUCTION Grid Computing Values and Risks CS693 Grid Computing Unit - I UNIT- I INTRODUCTION Grid Computing values and risks – History of Grid computing – Grid computing model and protocols – overview of types of Grids Grid Computing values and risks Key value elements that provide credible inputs to the various valuation metrics that can be used by enterprises to build successful Grid Computing deployment business cases. Each of the value elements can be applied to one, two, or all of the valuation models that a company may consider using, such as return on investment (ROI), total cost of ownership (TCO), and return on assets (ROA) Grid Computing Value Elements • Leveraging Existing Hardware Investments and Resources • Reducing Operational Expenses • Creating a Scalable and Flexible Enterprise IT Infrastructure • Accelerating Product Development, Improving Time to Market, and Raising Customer Satisfaction • Increasing Productivity Leveraging Existing Hardware Investments and Resources • Grids can be deployed on an enterprise’s existing infrastructure o mitigate the need for investment in new hardware systems. • eliminating expenditure on air conditioning, electricity, and in many cases, development of new data centers Example: grids deployed on existing desktops and servers provide over 93 percent in up-front hardware cost savings when compared to High Performance Computing Systems (HPC) Reducing Operational Expenses • Key self-healing and self-optimizing capabilities free system administrators from routine tasks o allow them to focus on high-value, important system administration • The operational expenses of a Grid Computing deployment are 73 percent less than for comparable HPC- based solutions • grids are being deployed in enterprises in as quickly as two days, with little or no disruption to operations • Cluster system deployments, on the other hand, are taking 60–90 days, in addition to the days required to configure and deploy the applications Creating a Scalable and Flexible Enterprise IT Infrastructure • Grid Computing allows companies to add resources linearly based on real-time business requirements • These resources can be derived from within the enterprise or from utility computing services • projects never have to be put on hold for a lack of computational capacity, space, or system priority • The entire compute infrastructure of the enterprise is available for connecting • Grid Computing can help bring about the end of departmental silos and expose computational assets curtained by server huggers and bureaucracy Accelerating Product Development, Improving Time to Market, and Raising Customer Satisfaction • The dramatic reduction in, simulation times can get products completed quickly. • This also provides the capability to perform a lot more detailed and exhaustive product design, as the computational resources brought to bear by the grid can quickly churn through the complex models and scenarios to detect design flaws Example: Life Sciences: Drug Discovery • introducing a “New Chemical Entity” (drug)” into the market costs US $802M @ 12–15 years MTech CSE (PT, 2011-14) SRM, Ramapuram 1 hcr:innovationcse@gg CS693 Grid Computing Unit - I • Grid Computing is allowing drug companies to get the most out of their R&D expenditure by developing the right product and getting it to market in the shortest possible time. • can save almost US $5M per month in R&D expenses of the drug development process • can amount to almost US $1M per day for each day that the product is brought to market early Increasing Productivity • run times of jobs submitted by its engineers were reduced by 58 percent by deploying a grid Example: Risk Analysis the key risk factors that plague technology deployments and analyze its vulnerabilities Lock-in • the IT manager will not be making any investments in durable complementary assets which promote lock-in • software, and supporting infrastructure and may contribute to customer lock-in. • pay keen attention to which vendors are supporting the Grid Computing standards at the Global Grid Forum Switching Costs • the primary switching cost will be driven by the effort required to integrate and enable enterprise applications to work on whatever replacement grid infrastructure that has been selected. • performed by utilizing software development toolkits • introduce new grid software in the enterprise to support new grid-enabled applications, while letting the existing software deployment and its integration with legacy grid software remain unchanged. Project Implementation Failure • project failure, either due to bad project management or incorrect needs assessment • take advantage of hosted pilot and professional services offered by grid software vendors • This will allow the IT manager to accurately pre-assess the suitability of the grid software, level of integration required, and feasibility (application speedup times, productivity gains, etc.). • Hosted pilots are conducted on the vendors’ data centers and have no impact to the company MTech CSE (PT, 2011-14) SRM, Ramapuram 2 hcr:innovationcse@gg CS693 Grid Computing Unit - I History of Grid Computing Academic Research Projects I-Way • 11 high speed networks were used to connect 17 sites with high-end computing resources for a demonstration to create one super “metacomputer* • Sixty different applications, spanning various faculties of science and engineering, were developed and run over this demonstration network. • Many of the early Grid Computing concepts were explored Globus • a suite of tools that laid the foundation for Grid Computing activities • 80 sites worldwide running software based on the Globus Toolkit were connected together Entropia • to harness the idle computers worldwide to solve problems of scientific interest • grew to 30,000 computers with aggregate speed of over one teraflop per second • ordinary users volunteered their PCs to analyze research topics such as patient’s response to chemotherapy, discovering drugs for AIDS, and potential cures for anthrax High-Performance Computing • refers to supercomputing • There are hundreds of supercomputers deployed throughout the world. • Key parallel processing algorithms have already been developed to support execution of programs on different, but co-located processors. • High-performance computing system deployment, is not limited to academic or research institutions. • The industries in which high performance systems are deployed are numerous in nature. • Example: Telecommunication, Finance, Automotive, Database, Transportation, Electronics, Geophysics, Aerospace, Energy, World Wide Web, Information Services, Chemistry, Manufacturing, Mechanics, Pharmaceutics Cluster Computing • Clusters are high-performance, massively parallel computers built primarily out of commodity hardware components, running a free-software operating system such as Linux or FreeBSD, and interconnected by a private high-speed network. • It consists of a cluster of PCs, or workstations, dedicated to running high-performance computing tasks. • The nodes in the cluster do not sit on users’ desks, but are dedicated to running cluster jobs. • A cluster is usually connected to the outside world through only a single node • numerous tools have been developed to run and manage clusters • load-balancing tools, tools to adapt applications run in the parallel cluster environment (Ex: ForgeExplorer) MTech CSE (PT, 2011-14) SRM, Ramapuram 3 hcr:innovationcse@gg CS693 Grid Computing Unit - I Peer-to-Peer Computing • Two main models: centralized model, decentralized model Centralized model • file sharing is based around the use of a central server system that directs traffic between individual registered users. • The central servers maintain directories of the shared files stored on the respective PCs of registered users of the network. • These directories are updated every time a user logs on or off the Napster server network. • Each time a user of a centralized P2P file sharing system submits a request or searches for a particular file, the central server creates a list of files matching the search request by cross-checking the request with the server’s database of files belonging to users who are currently connected to the network. • The central server then displays that list to the requesting user. • The requesting user can then select the desired file from the list and open a direct HTTP link with the individual computer that currently possesses that file. • The download of the actual file takes place directly, from one network user to the other. • The actual file is never stored on the central server or on any intermediate point on the network. Decentralized model • file sharing does not use a central server to keep track of files. • it relies on each individual computer to announce its existence to a peer, which in turn announces it to all the users that it is connected to, and so on. • If one of the computers in the peer network has a file that matches the request, it transmits the file information (name, size) back through all the computers in the pathway to the user that requested the file. • A direct connection is established and the file is transferred Internet Computing • utilize the vast processing cycles available at users’ desktops • Large compute intensive projects are coded so that tasks can be broken down into smaller subtasks and distributed over the Internet for processing. • Volunteer users then download a lightweight client onto their desktop, which periodically communicates with the central server to receive tasks. • The client initiates
Recommended publications
  • Job Peer Group Controller Job Peer Group Host „ a PG for Each Job
    Korea-Japan Grid Symposium, Seoul December 1-2, 2004 P3: Personal Power Plant Makes over your PCs into power generator on the Grid Kazuyuki Shudo <[email protected]>, Yoshio Tanaka, Satoshi Sekiguchi Grid Technology Research Center, AIST, Japan P3: Personal Power Plant Middleware for distributed computation utilizing JXTA. Traditional goals Cycle scavenging Harvest compute power of existing PCs in an organization. Conventional dist. computing Internet-wide distributed computing E.g. distributed.net, SETI@home Challenging goals Aggregate PCs and expose them as an integrated Grid resource. Integrate P3 with Grid middleware ? cf. Community Scheduler Framework Dealings and circulation of computational resources Transfer individual resources (C2C, C2B) and also aggregated resources (B2B). Transfer and aggregation of Other resources than processing power. individual resources Commercial dealings need a market and a system supporting it. P2P way of interaction between PCs P3 uses JXTA for all communications JXTA is a widely accepted P2P protocol, project and library, that provides common functions P2P software requires. P2P concepts supported by JXTA efficiently support P3: Ad-hoc self-organization PCs can discover and communicate with each other without pre-configuration. Discovery PCs dynamically discover each other and jobs without a central server. Peer Group PCs are grouped into job groups, in which PCs carry out code distribution, job control, and collective communication for parallel computation. Overlay Network Peer ID in JXTA is independent from physical IDs like IP addresses and MAC addresses. JXTA enables end-to-end bidirectional communication over NA(P)T and firewall (even if the FW allows only unidirectional HTTP). This function supports parallel processing in the message-passing model, not only master-worker model.
    [Show full text]
  • Web Services and PC Grid
    Web Services and PC Grid ONG Guan Sin <[email protected]> Grid Evangelist Singapore Computer Systems Ltd APBioNet/APAN 18 Jul 2006 Tera-scale Campus Grid @ NUS LATEST: CIO Award 2006 winner Harnessing existing computation capacity campus-wide, creating large-scale supercomputing capability Computers are aggregated through its gigabit network into a virtual supercomputing platform using United Devices Grid MP middleware Community grid by voluntary participation from depts Two-year collaboration project between NUS and SCS to develop applications and support community Number of Nodes Theoretical# Practical^ 1,042 (Sep 20, 2005) 4.5 TFlops 2.5 TFlops 3,000 (Planned - 2007) 13 TFlops 7.2 TFlops * Accumulated average CPU speed of 2.456GHz as at Sep 20, 2005 # Assuming 90% capacity effectively harnessed ^ Assuming 50% capacity effectively harnessed Copyright 2006 Singapore Computer Systems Limited 2 UD Grid MP Architecture Managed Grid Services Interface (MGSI) ± Web Services API 3 Command-line Interface Application Service 4 Simple Web Interface 5 Application Service Overview Application Service . Is a job submission and result retrieval program which provides users with a simple interface for performing work on the Grid . It is responsible for Splitting and Merging Application Data Features . Control the Grid MP Sever with SOAP or XML-RPC communications . SOAP and XML-RPC Communications are protected with SSL encryption . SOAP and XML-RPC are language and platform independent with many publicly available toolkits and libraries. User interface can be command-line, web-based, GUI, etc. Can be written to run on various operating systems MP Grid Services Interface (MGSI) . All Objects in the Grid MP platform can be controlled through the MGSI .
    [Show full text]
  • Good Practice Guide No. 17 Distributed Computing for Metrology
    NPL REPORT DEM-ES 006 Software Support for Metrology - Good Practice Guide No. 17 Distributed computing for metrology applications T J Esward, N J McCormick, K M Lawrence and M J Stevens NOT RESTRICTED March 2007 National Physical Laboratory | Hampton Road | Teddington | Middlesex | United Kingdom | TW11 0LW Switchboard 020 8977 3222 | NPL Helpline 020 8943 6880 | Fax 020 8943 6458 | www.npl.co.uk Software Support for Metrology Good Practice Guide No. 17 Distributed computing for metrology applications T J Esward, N J McCormick, K M Lawrence and M J Stevens Mathematics and Scientific Computing Group March 2007 ABSTRACT This guide aims to facilitate the effective use of distributed computing methods and techniques by other National Measurement System (NMS) programmes and by metrologists in general. It focuses on the needs of those developing applications for distributed computing systems, and on PC desktop grids in particular, and seeks to ensure that application developers are given enough knowledge of system issues to be able to appreciate what is needed for the optimum performance of their own applications. Within metrology, the use of more comprehensive and realistic mathematical models that require extensive computing resources for their solution is increasing. The motivation for the guide is that in several areas of metrology the computational requirements of such models are so demanding that there is a strong requirement for distributed processing using parallel computing on PC networks. Those who need to use such technology can benefit from guidance on how best to structure the models and software to allow the effective use of distributed computing.
    [Show full text]
  • Print This Article
    L{{b b Volume 1, No. 3, Sept-Oct 2010 International Journal of Advanced Research in Computer Science RESEARCH PAPER Available Online at www.ijarcs.info A Simplified Network manager for Grid and Presenting the Grid as a Computation Providing Cloud M.Sudha* M.Monica Assistant Professor (Senior) Assistant Professor, School of Information Technology and Engineering, VIT School of Computer Science and Engineering, University INDIA VIT University INDIA [email protected] [email protected] Abstract: One of the common forms of distributed computing is grid computing. A grid uses the resources of many separate computers, loosely connected by a network, to solve large-scale computation problems. Our approach was as follows first computationally large data is split into a number of smaller, more manageable, working units secondly each work-unit is then sent to one member of the grid ,That member completes processing of that work-unit in its own and sends back the result. In this architecture, there needs to be at least one host that performs the task of assigning work-units, and then sending them, to a remote processor, as well as receive the results from remote processors. We call this unit as the Network Manager. In addition to this assigning, sending and receiving the work-units and results, there also is the need for a host that splits tasks into work-units and assimilates the received work units. We call this unit as the Task Broker, which we propose to design. On the server end, there is a program for processing module, splitter and assimilator (broker).
    [Show full text]
  • ETSI TR 102 659-1 V1.2.1 (2009-10) Technical Report
    ETSI TR 102 659-1 V1.2.1 (2009-10) Technical Report GRID; Study of ICT Grid interoperability gaps; Part 1: Inventory of ICT Stakeholders 2 ETSI TR 102 659-1 V1.2.1 (2009-10) Reference RTR/GRID-0001-1[2] Keywords analysis, directory, ICT, interoperability, testing ETSI 650 Route des Lucioles F-06921 Sophia Antipolis Cedex - FRANCE Tel.: +33 4 92 94 42 00 Fax: +33 4 93 65 47 16 Siret N° 348 623 562 00017 - NAF 742 C Association à but non lucratif enregistrée à la Sous-Préfecture de Grasse (06) N° 7803/88 Important notice Individual copies of the present document can be downloaded from: http://www.etsi.org The present document may be made available in more than one electronic version or in print. In any case of existing or perceived difference in contents between such versions, the reference version is the Portable Document Format (PDF). In case of dispute, the reference shall be the printing on ETSI printers of the PDF version kept on a specific network drive within ETSI Secretariat. Users of the present document should be aware that the document may be subject to revision or change of status. Information on the current status of this and other ETSI documents is available at http://portal.etsi.org/tb/status/status.asp If you find errors in the present document, please send your comment to one of the following services: http://portal.etsi.org/chaircor/ETSI_support.asp Copyright Notification No part may be reproduced except as authorized by written permission. The copyright and the foregoing restriction extend to reproduction in all media.
    [Show full text]
  • Resume Ivo Janssen
    Ivo J. Janssen 13266 Kerrville Folkway Austin, TX 78729 (512) 750-9455 / [email protected] Objective To be part of a successful professional services team in a pre-sales and post-sales architecture and deployment role. Summary I'm an experienced, enthusiastic and skilled professional services consultant with experience in architecting and supporting large-scale enterprise software, with extensive experience in designing and implementing custom solutions as part of pre- and post sales support activities in diverse industry verticals across five continents. I'm as comfortable in designing and coding in an engineering role as I am in on-site planning, executing and troubleshooting customer deployments in a professional services role, as well as in participating in sales calls and drafting proposals in a sales engineering role. I'm able to adapt to new environments quickly and rapidly become proficient with new systems, tools and skills. I'm a good team player who thinks beyond the problem at hand instead of following the established paths in order to find a better solution that meets both current and future needs. Professional experience • Virtual Bridges , Austin, TX, USA (Nov 2010 – present) Senior Solution Architect o Solution architect for VDI product line. • Initiate Systems / IBM , Austin, TX, USA (Sep 2008 – Nov 2010) Senior Consultant o Senior implementation engineer for Healthcare sector of Master Data Management product line. o Responsibilities include requirements gathering, product installation and configuration, custom coding, connectivity scripting on Windows and Unix, database performance tuning, customer training workshops, pre/post-sales support. o Successfully architected, implemented and brought live custom implementation projects for major hospital systems and labs in the US, including customers with up to 250 million records (1TB database) connected to 20 auxiliary systems.
    [Show full text]
  • Desktop Grids for Escience
    Produced by the IDGF-SP project for the International Desktop Grid Federation A Road Map Desktop Grids for eScience Technical part – December 2013 IDGF/IDGF-SP International Desktop Grid federation http://desktopgridfederation.org Edited by Ad Emmen Leslie Versweyveld Contributions Robert Lovas Bernhard Schott Erika Swiderski Peter Hannape Graphics are produced by the projects. version 4.2 2013-12-27 © 2013 IDGF-SP Consortium: http://idgf-sp.eu IDGF-SP is supported by the FP7 Capacities Programme under contract nr RI-312297. Copyright (c) 2013. Members of IDGF-SP consortium, see http://degisco.eu/partners for details on the copyright holders. You are permitted to copy and distribute verbatim copies of this document containing this copyright notice but modifying this document is not allowed. You are permitted to copy this document in whole or in part into other documents if you attach the following reference to the copied elements: ‘Copyright (c) 2013. Members of IDGF-SP consortium - http://idgf-sp.eu’. The commercial use of any information contained in this document may require a license from the proprietor of that information. The IDGF-SP consortium members do not warrant that the information contained in the deliverable is capable of use, or that use of the information is free from risk, and accept no liability for loss or damage suffered by any person and organisation using this information. – 2 – Preface This document is produced by the IDGF-SP project for the International Desktop Grid Fe- deration. Please note that there are some links in this document pointing to the Desktop Grid Federation portal, that requires you to be signed in first.
    [Show full text]
  • Distributed Computing with the Berkeley Open Infrastructure for Network Computing BOINC
    Distributed Computing with the Berkeley Open Infrastructure for Network Computing BOINC Eric Myers 1 September 2010 Mid-Hudson Linux Users Group 2 How BOINC Works BOINC Client BOINC is the software BOINC Server Windows framework that makes Linux Mac OS this all work. Linux 50+ separate projects (& Solaris, AIX, HP-UX, etc…) 1 September 2010 Mid-Hudson Valley Linux Users Group 3 BOINC Dataflow 1 September 2010 Mid-Hudson Valley Linux Users Group 4 Early History SETI@home . May 1999 to Dec 2005 (“Classic”) . 2003 (BOINC) World Community Grid (IBM) . Nov 2004 to mid 2008 (Grid MP) Climatepredition.net . Nov 2007 (BOINC) . Sept 2003 (“Classic”) . August 2004 (BOINC) Einstein@Home . February 2005 Predictor@Home . (Pirates@Home – June 2004 :-) . June 2004 - Scripps . October 2008 – U. Michigan Rosetta@Home (University of Washington) . June 2005 LHC@Home . Sept 2004 - CERN PrimeGrid (Lithuania) . October 2007 - QMC . July 2005 With 50+ to follow… 1 September 2010 Mid-Hudson Valley Linux Users Group 5 http://setiathome.berkeley.edu SETI@home http:/climateprediction.net/ 1 September 2010 Mid-Hudson Valley Linux Users Group 7 http://www.worldcommunitygrid.org/ World Community Grid Active The Clean Energy Project - Phase 2 Help Cure Muscular Dystrophy – Phase 2 Funded and operated by IBM Help Fight Childhood Cancer Help Conquer Cancer Human Proteome Folding - Phase 2 Completed FightAIDS@Home Nutritious Rice for the World Intermittent AfricanClimate@Home Discovering Dengue Drugs - Together - Phase 2 Help Cure Muscular Dystrophy Influenza Antiviral Drug Search Genome Comparison The Clean Energy Project Help Defeat Cancer Discovering Dengue Drugs - Together Human Proteome Folding 1 September 2010 Mid-Hudson Valley Linux Users Group 8 http://einstein.phys.uwm.edu/ or http://einsteinathome.org Einstein@Home 9 Rules and Policies Run BOINC only on authorized computers Run BOINC only on computers that you own, or for which you have obtained the owner’s permission.
    [Show full text]
  • CS693 Unit V
    CS693 – Grid Computing Unit - V UNIT – V APPLICATIONS, SERVICES AND ENVIRONMENTS Application integration – application classification – Grid requirements – Integrating applications with Middleware platforms – Grid enabling Network services – managing Grid environments – Managing Grids – Management reporting – Monitoring – Data catalogs and replica management – portals – Different application areas of Grid computing Application Classification The dimensions relevant to enabling the application to run on a grid are o parallelism o granularity o communications o dependency affects the way in which it can be migrated to a grid environment not mutually exclusive—a specific application can be characterized along most, of these dimensions. Parallelism classification scheme attributed to Flynn has a significant impact on how the application is integrated to run on a grid. Single Program, Single Data (SPSD) simple sequential programs that take a single input set and generate a single output set Motivations 1) A vast number of computing resources are immediately available grid is used as a throughput engine when many instances of this single program need to be executed. Leveraging resources outside the local domain can vastly improve the throughput of these jobs. Example: Logic simulation of a microprocessor design The design needs to be verified by running millions of test cases against the design. By leveraging grid resources, each test case can be run on a different remote machine, thereby improving the overall throughput of the simulation. 2) Remotely available shared data can be used program can be made to execute where the data reside or the data can be accessed via a “data grid.” grid environment enables controlled access to shared data.
    [Show full text]
  • The Organic Grid: Self-Organizing Computation on a Peer-To-Peer Network
    The Organic Grid: Self-Organizing Computation on a Peer-to-Peer Network Arjav J. Chakravarti Gerald Baumgartner Mario Lauria Dept. of Computer and Information Science The Ohio State University 395 Dreese Lab., 2015 Neil Ave. Columbus, OH 43210–1277, USA Email: arjav, gb, lauria @cis.ohio-state.edu Technical Report: OSU-CISRC-10/03-TR55 October 24, 2003 Abstract Desktop grids have already been used to perform some of the largest computations in the world and have the potential to grow by several more orders of magnitude. However current approaches to utilizing desktop resources require either centralized servers or extensive knowledge of the underlying system, limiting their scalability. We propose a biologically inspired and fully-decentralized approach to the organization of compu- tation that is based on the autonomous scheduling of strongly mobile agents on a peer-to-peer network. In a radical departure from current models, we envision large scale desktop grids in which agents auto- nomically organize themselves so as to maximize resource utilization. We demonstrate this concept with a reduced scale proof-of-concept implementation that executes a data-intensive parameter sweep application on a set of heterogeneous geographically distributed ma- chines. We present a detailed exploration of the design space of our system and a performance evaluation of our implementation using metrics appropriate for assessing self-organizing desktop grids. 1 Introduction Some of the largest computations in the world have been carried out on collections of PCs and workstations over the Internet. Tera-flop levels of computational power have been achieved by systems composed of heterogeneous computing resources that number in the hundreds-of-thousands to the millions.
    [Show full text]
  • A .NET-Based Cross-Platform Software for Desktop Grids Alireza Poshtkohi
    313 DotGrid: a .NET-based cross-platform software for desktop grids Alireza Poshtkohi* Department of Electrical Engineering Qazvin Azad University Daneshgah St., P.O. Box 34185–1416, Qazvin, Iran E-mail: [email protected] *Corresponding author Ali Haj Abutalebi and Shaahin Hessabi Department of Computer Engineering Sharif University of Technology Azadi St., P.O. Box 11365–8639, Tehran, Iran E-mail: [email protected] E-mail: [email protected] Abstract: Grid infrastructures that have provided wide integrated use of resources are becoming the de facto computing platform for solving large-scale problems in science, engineering and commerce. In this evolution, desktop grid technologies allow the grid communities to exploit the idle cycles of pervasive desktop PC systems to increase the available computing power. In this paper we present DotGrid, a cross-platform grid software. DotGrid is the first comprehensive desktop grid software utilising Microsoft’s .NET Framework in Windows-based environments and MONO .NET in Unix-class operating systems to operate. Using DotGrid services and APIs, grid desktop middleware and applications can be implemented conveniently. We evaluated our DotGrid’s performance by implementing a set of grid-based applications. Keywords: desktop grids; grid computing; cross-platform grid software. Reference to this paper should be made as follows: Poshtkohi, A., Abutalebi, A.H. and Hessabi, S. (2007) ‘DotGrid: a .NET-based cross-platform software for desktop grids’, Int. J. Web and Grid Services, Vol. 3, No. 3, pp.313–332. Biographical notes: Alireza Poshtkohi received his BSc degree in Electrical Engineering from the Islamic Azad University of Qazvin, Qazvin, Iran in 2006.
    [Show full text]
  • Good Practice Guide No. 17 Distributed Computing for Metrology
    NPL REPORT DEM-ES 006 Software Support for Metrology - Good Practice Guide No. 17 Distributed computing for metrology applications T J Esward, N J McCormick, K M Lawrence and M J Stevens NOT RESTRICTED March 2007 National Physical Laboratory | Hampton Road | Teddington | Middlesex | United Kingdom | TW11 0LW Switchboard 020 8977 3222 | NPL Helpline 020 8943 6880 | Fax 020 8943 6458 | www.npl.co.uk Software Support for Metrology Good Practice Guide No. 17 Distributed computing for metrology applications T J Esward, N J McCormick, K M Lawrence and M J Stevens Mathematics and Scientific Computing Group March 2007 ABSTRACT This guide aims to facilitate the effective use of distributed computing methods and techniques by other National Measurement System (NMS) programmes and by metrologists in general. It focuses on the needs of those developing applications for distributed computing systems, and on PC desktop grids in particular, and seeks to ensure that application developers are given enough knowledge of system issues to be able to appreciate what is needed for the optimum performance of their own applications. Within metrology, the use of more comprehensive and realistic mathematical models that require extensive computing resources for their solution is increasing. The motivation for the guide is that in several areas of metrology the computational requirements of such models are so demanding that there is a strong requirement for distributed processing using parallel computing on PC networks. Those who need to use such technology can benefit from guidance on how best to structure the models and software to allow the effective use of distributed computing.
    [Show full text]