Libarcclient
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Grid Computing: What Is It, and Why Do I Care?*
Grid Computing: What Is It, and Why Do I Care?* Ken MacInnis <[email protected]> * Or, “Mi caja es su caja!” (c) Ken MacInnis 2004 1 Outline Introduction and Motivation Examples Architecture, Components, Tools Lessons Learned and The Future Questions? (c) Ken MacInnis 2004 2 What is “grid computing”? Many different definitions: Utility computing Cycles for sale Distributed computing distributed.net RC5, SETI@Home High-performance resource sharing Clusters, storage, visualization, networking “We will probably see the spread of ‘computer utilities’, which, like present electric and telephone utilities, will service individual homes and offices across the country.” Len Kleinrock (1969) The word “grid” doesn’t equal Grid Computing: Sun Grid Engine is a mere scheduler! (c) Ken MacInnis 2004 3 Better definitions: Common protocols allowing large problems to be solved in a distributed multi-resource multi-user environment. “A computational grid is a hardware and software infrastructure that provides dependable, consistent, pervasive, and inexpensive access to high-end computational capabilities.” Kesselman & Foster (1998) “…coordinated resource sharing and problem solving in dynamic, multi- institutional virtual organizations.” Kesselman, Foster, Tuecke (2000) (c) Ken MacInnis 2004 4 New Challenges for Computing Grid computing evolved out of a need to share resources Flexible, ever-changing “virtual organizations” High-energy physics, astronomy, more Differing site policies with common needs Disparate computing needs -
The Translational Journey of the Htcondor-CE
Journal of Computational Science xxx (xxxx) xxx Contents lists available at ScienceDirect Journal of Computational Science journal homepage: www.elsevier.com/locate/jocs Principles, technologies, and time: The translational journey of the HTCondor-CE Brian Bockelman a,*, Miron Livny a,b, Brian Lin b, Francesco Prelz c a Morgridge Institute for Research, Madison, USA b Department of Computer Sciences, University of Wisconsin-Madison, Madison, USA c INFN Milan, Milan, Italy ARTICLE INFO ABSTRACT Keywords: Mechanisms for remote execution of computational tasks enable a distributed system to effectively utilize all Distributed high throughput computing available resources. This ability is essential to attaining the objectives of high availability, system reliability, and High throughput computing graceful degradation and directly contribute to flexibility, adaptability, and incremental growth. As part of a Translational computing national fabric of Distributed High Throughput Computing (dHTC) services, remote execution is a cornerstone of Distributed computing the Open Science Grid (OSG) Compute Federation. Most of the organizations that harness the computing capacity provided by the OSG also deploy HTCondor pools on resources acquired from the OSG. The HTCondor Compute Entrypoint (CE) facilitates the remote acquisition of resources by all organizations. The HTCondor-CE is the product of a most recent translational cycle that is part of a multidecade translational process. The process is rooted in a partnership, between members of the High Energy Physics community and computer scientists, that evolved over three decades and involved testing and evaluation with active users and production infrastructures. Through several translational cycles that involved researchers from different organizations and continents, principles, ideas, frameworks and technologies were translated into a widely adopted software artifact that isresponsible for provisioning of approximately 9 million core hours per day across 170 endpoints. -
GT 4.0 RLS GT 4.0 RLS Table of Contents
GT 4.0 RLS GT 4.0 RLS Table of Contents 1. Key Concepts .................................................................................................................................. 1 1. Overview of Data Management in GT4 ........................................................................................ 1 2. Data movement ........................................................................................................................ 1 3. Data replication ....................................................................................................................... 2 4. Higher level data services .......................................................................................................... 4 2. 4.0.0 Release Note ............................................................................................................................ 5 1. Component Overview ............................................................................................................... 5 2. Feature Summary ..................................................................................................................... 5 3. Bug Fixes ............................................................................................................................... 5 4. Known Problems ...................................................................................................................... 6 5. Technology Dependencies ......................................................................................................... -
A Globus Primer Or, Everything You Wanted to Know About Globus, but Were Afraid to Ask Describing Globus Toolkit Version 4
A Globus Primer Or, Everything You Wanted to Know about Globus, but Were Afraid To Ask Describing Globus Toolkit Version 4 An Early and Incomplete Draft Please send comments, criticisms, and suggestions to: [email protected] Preface The Globus Toolkit (GT) has been developed since the late 1990s to support the development of service-oriented distributed computing applications and infrastructures. Core GT components address basic issues relating to security, resource access and management, data movement and management, resource discovery, and so forth. A broader “Globus universe” comprises numerous tools and components that build on core GT4 functionality to provide many useful application- level functions. These tools have been used to develop many Grid systems and applications. Version 4 of the Globus Toolkit, GT4, released in April 2005, represents a significant advance relative to the GT3 implementation of Web services functionality in terms of the range of components provided, functionality, standards conformance, usability, and quality of documentation. This document is intended to provide a first introduction to key features of both GT4 and associated tools, and the ways in which these components can be used to develop Grid infrastructures and applications. Its focus is on the user’s view of the technology and its application, and the practical techniques that should be employed to develop GT4-based applications. We discuss in turn the applications that motivate the development of GT4 and related tools; the four tasks involved in building Grids: design, deployment, application, operations; GT4 structure, including its Web services (WS) and pre-WS components; the Globus universe and its various components; GT4 performance, scalability, and functionality testing; porting to GT4 from earlier versions of GT; who’s using GT4; likely future directions, and a few words of history. -
Parallelism of Machine Learning Algorithms
10/9/2013 Outline Condor Presented by : Walid Budgaga The Anatomy of the Grid Globus Toolkit CS 655 – Advanced Topics in Distributed Systems Computer Science Department Colorado State University 1 2 Motivation Motivation High Throughput Computing (HTC)? HTC is suitable for scientific research Large amounts of computing capacity over long Example(Parameter sweep): periods of time. Testing parameter combinations to Measured: operations per month or per year High Performance Computing (HPC)? keep temp. at particular level Large amounts of computing capacity for short periods of time op(x,y,z) takes 10 hours, 500(MB) memory, I/O 100(MB) Measured: FLOPS x(100), y(50), z(25)=> 100x50x25=125,000(145 years) 3 4 Motivation HTC Environment Fort Collins Large amounts of processing capacity? Exploiting computers on the network Science Center Utilizing heterogeneous resources Uses Condor for Overcoming differences of the platforms By building portable solution scientific projects Including resource management framework Over long periods of time? System must be reliable and maintainable Surviving failures (software & hardware) Allowing leaving and joining of resources at any time Upgrading and configuring without significant downtimes Source: http://www.fort.usgs.gov/Condor/ComputingTimes.asp 5 6 1 10/9/2013 HTC Environment HTC Also, the system must meet the needs of: Other considerations: Resource owners The distributive owned resources lead to: Rights respected Decentralized maintenance and configuration of resources Policies enforced Resource availability Customers Benefit of additional processing capacity outweigh complexity of Applications preempted at any time usage Adds an additional degree of resource heterogeneity System administrators Real benefit provided to users outweigh the maintenance cost 7 8 Condor Overview Open-source high-throughput computing framework for computing intensive tasks. -
The Advanced Resource Connector User Guide
NORDUGRID NORDUGRID-MANUAL-6 7/12/2010 The NorduGrid/ARC User Guide Advanced Resource Connector (ARC) usage manual 2 Contents 1 Preface 11 2 Roadmap 13 3 Client Installation 15 3.1 Beginner Installation . 15 3.1.1 Download . 15 3.1.2 Unpack . 15 3.1.3 Configure . 16 3.2 System-wide installation . 16 3.2.1 Installation via repositories . 17 3.2.2 Globus Toolkit . 18 3.2.3 Download the client . 18 3.2.4 Build . 19 3.2.5 Install the client . 20 3.3 Client configuration files . 21 3.3.1 client.conf . 21 3.3.2 srms.conf . 23 3.3.3 Deprecated configuration files . 24 4 Grid Certificates 25 4.1 Quick start with certificates . 25 4.1.1 Certificates in local standalone client . 25 4.1.2 Certificates for system-wide client installation . 27 4.1.3 Obtain a personal certificate . 27 4.2 Grid Authentication And Certificate Authorities . 28 4.2.1 Grid Login, Proxies . 28 4.2.2 Certificate Authorities . 29 4.2.3 Globus Grid certificates . 29 4.2.4 Associating yourself with a proper CA . 30 4.2.5 Friendly CAs . 31 4.2.6 Certificate Request . 32 4.2.7 Working with certificates: examples . 33 3 4 CONTENTS 5 Getting Access to Grid Resources: Virtual Organisations 35 5.1 NorduGrid Virtual Organisation . 35 5.2 Other Virtual Organisations . 36 6 Grid Session 37 6.1 Logging Into The Grid . 37 6.1.1 Proxy Handling Tips . 38 6.2 First Grid test . 38 6.3 Logging Out . -
Facilitating E-Science Discovery Using Scientific Workflows on the Grid
Chapter 13 Facilitating e-Science Discovery Using Scientific Workflows on the Grid Jianwu Wang, Prakashan Korambath, Seonah Kim, Scott Johnson, Kejian Jin, Daniel Crawl, Ilkay Altintas, Shava Smallen, Bill Labate, and Kendall N. Houk Abstract e-Science has been greatly enhanced from the developing capability and usability of cyberinfrastructure. This chapter explains how scientific workflow systems can facilitate e-Science discovery in Grid environments by providing features including scientific process automation, resource consolidation, parallelism, provenance tracking, fault tolerance, and workflow reuse. We first overview the core services to support e-Science discovery. To demonstrate how these services can be seamlessly assembled, an open source scientific workflow system, called Kepler, is integrated into the University of California Grid. This architecture is being applied to a computational enzyme design process, which is a formidable and collaborative problem in computational chemistry that challenges our knowledge of protein chemistry. Our implementation and experiments validate how the Kepler workflow system can make the scientific computation process automated, pipe- lined, efficient, extensible, stable, and easy-to-use. 13.1 Introduction “e-Science is about global collaboration in key areas of science and the next genera- tion of infrastructure that will enable it.”1 Grid computing “coordinates resources that are not subject to centralized control by using standard, open, general-purpose 1 John Taylor, Director General of Research Councils, Office of Science and Technology, UK. J. Wang (*), D. Crawl, I. Altintas, and S. Smallen San Diego Supercomputer Center, UCSD, 9500 Gilman Drive, MC 0505, La Jolla, CA 92093, USA e-mail: [email protected] P. Korambath, K. -
Reflections on Building a High-Performance Computing Cluster Using Freebsd
Reflections on Building a High-performance Computing Cluster Using FreeBSD NYCBUG March 2008 Brooks Davis The Aerospace Corporation El Segundo, California, USA http://people.freebsd.org/~brooks/pubs/meetbsd2007/ © Copyright 2007 The Aerospace Corporation Outline ● Fellowship Overview ● Evaluation of Design Issues ● Lessons Learned ● Thoughts on Future Clusters ● Conclusions A Brief History of Fellowship ● Started in 2001 ● Primarily motivated by the needs of the GPS program office ● Intended to be a corporate resource – Diverse set of users Fellowship Hardware ● 352 dual-processor nodes – 64 Quad-core Woodcrest Xeons – 288 Opterons (152 dual-core) – 1-4GB RAM, 80-250GB disk ● 6 core systems – fellowship – shell server – fellowship64 – amd64 shell server – arwen – node netboot server, scheduler, NIS, DNS, Mathematica License Server – elrond – /scratch – moria – NetApp FAS250 /home, /usr/aero Fellowship Circa February, 2007 Fellowship Composition by processor count Fellowship Composition by core count Fellowship Network r01n01 r01n02 fellowship r01n03 ... frodo r02n01 10.5.0.0/16 Aerospace gamgee r02n02 r02n03 Network arwen Cat6509 ... elrond r03n01 r03n01 r03n02 moria r03n02 r03n03 r03n03 ... ... Fellowship Software ● FreeBSD 6.2-RELEASE-p8 ● Sun Grid Engine (SGE) 6.0u11 ● Message Passing Interface (MPI) – MPICH – MPICH2 – OpenMPI ● PVM ● Mathematica ● Matlab Design Issues ● The Issue ● What We Did ● How it worked Operating System ● FreeBSD 4.x initially – Many years of service – Userland threading forced some application design choices -
Glite Middleware
Enabling Grids for E-sciencE Overview of the EGEE project and the gLite middleware www. eu-egee.org EGEE-III INFSO-RI-222667 Outline Enabling Grids for E-sciencE • What is EGEE? – The project – The infrastructure • gLite middleware • EGEE applications • Sources of further information EGEE-III INFSO-RI-222667 2 Defining the Grid Enabling Grids for E-sciencE • A Grid is the combination of networked resources and the corresponding middleware, which provides services for the user. EGEE-III INFSO-RI-222667 Providing a Production Grid Infrastructure for Collaborative Science 3 The EGEE Project Enabling Grids for E-sciencE • Aim of EGEE: “to establish a seamless European Grid infrastructure for the support of the European Research Area (ERA)” • EGEE – 1 April 2004 – 31 March 2006 – 71 partners in 27 countries, federated in regional Grids • EGEE-II – 1 April 2006 – 30 April 2008 – EddtiExpanded consortium • EGEE-III – 1 May 2008 – 30 April 2010 – Transition to sustainable model EGEE-III INFSO-RI-222667 Providing a Production Grid Infrastructure for Collaborative Science 4 Defining the Grid Enabling Grids for E-sciencE • A Grid is the combination of networked resources and the corresponding middleware, which provides services for the user. EGEE-III INFSO-RI-222667 Providing a Production Grid Infrastructure for Collaborative Science 5 EGEE working with related Enabling Grids for E-sciencE infrastructure projects GIN EGEE-III INFSO-RI-222667 Providing a Production Grid Infrastructure for Collaborative Science 6 What is happening now? Enabling Grids -
Pos(EGICF12-EMITC2)050
A Science Gateway Getting Ready for Serving the International Molecular Simulation Community PoS(EGICF12-EMITC2)050 Sandra Gesing Center for Bioinformatics & Department of Computer Science, University of Tübingen Sand 14, 72076 Tübingen, Germany E-mail: [email protected] Sonja Herres-Pawlis1 Department of Chemistry, Ludwig-Maximilians-University Munich Butenandtstr. 5-13, 81377 München, Germany E-mail: [email protected] Georg Birkenheuer Paderborn Center for Parallel Computing, University of Paderborn Warburger Str. 100, 33089 Paderborn, Germany E-mail: [email protected] André Brinkmann Johannes Gutenberg-University Mainz 55099 Mainz, Germany E-mail: [email protected] Richard Grunzke Center for Information Services and High Performance Computing, Technische Universität Dresden Zellescher Weg 12-14, Germany E-mail: [email protected] Peter Kacsuk Laboratory of Parallel and Distributed Systems, MTA SZTAKI Kende Street 13-17, 1111 Budapest, Hungary E-mail: [email protected] Oliver Kohlbacher Center for Bioinformatics & Department of Computer Science, University of Tübingen Sand 14, 72076 Tübingen, Germany E-mail: [email protected] 1 Speaker Copyright owned by the author(s) under the terms of the Creative Commons Attribution-NonCommercial-ShareAlike Licence. http://pos.sissa.it Miklos Kozlovszky Laboratory of Parallel and Distributed Systems, MTA SZTAKI Kende Street 13-17, 1111 Budapest, Hungary E-mail: [email protected] Jens Krüger Center for Bioinformatics & Department -
Experimental Study of Remote Job Submission and Execution on LRM Through Grid Computing Mechanisms
Experimental Study of Remote Job Submission and Execution on LRM through Grid Computing Mechanisms Harshadkumar B. Prajapati Vipul A. Shah Information Technology Department Instrumentation & Control Engineering Dharmsinh Desai University Department Nadiad, INDIA Dharmsinh Desai University e-mail: [email protected], Nadiad, INDIA [email protected] e-mail: [email protected], [email protected] Abstract —Remote job submission and execution is The fundamental unit of work-done in Grid fundamental requirement of distributed computing done computing is successful execution of a job. A job in using Cluster computing. However, Cluster computing Grid is generally a batch-job, which does not interact limits usage within a single organization. Grid computing with user while it is running. Higher-level Grid environment can allow use of resources for remote job services and applications use job submission and execution that are available in other organizations. This execution facilities that are supported by a Grid paper discusses concepts of batch-job execution using LRM and using Grid. The paper discusses two ways of middleware. Therefore, it is very important to preparing test Grid computing environment that we use understand how a job is submitted, executed, and for experimental testing of concepts. This paper presents monitored in Grid computing environment. experimental testing of remote job submission and Researchers and scientists who are working in higher- execution mechanisms through LRM specific way and level Grid services and applications do not pay much Grid computing ways. Moreover, the paper also discusses attention to the underlying Grid infrastructure in their various problems faced while working with Grid published work. -
Introduction to Grid Computing with the Globus Toolkit Version 3 Page 1 Colorado Software Summit: October 26 – 31, 2003 © Copyright 2003, IBM Corporation
Colorado Software Summit: October 26 – 31, 2003 © Copyright 2003, IBM Corporation Introduction to Grid Programming with the Globus Toolkit Version 3 Michael Brown IBM Linux Integration Center Austin, Texas Michael Brown – Introduction to Grid Computing with the Globus Toolkit Version 3 Page 1 Colorado Software Summit: October 26 – 31, 2003 © Copyright 2003, IBM Corporation Session Introduction Who am I? ➢ [email protected] ➢ Team Leader for Americas Projects ➢ IBM Linux Integration Center, Austin ➢ Author on 2 IBM Redbooks about Globus 3 What will we talk about ? ➢ Recent history of grid computing ➢ Intro to the galaxy of Globus ➢ Run a Globus V3 example (fingers crossed) Michael Brown – Introduction to Grid Computing with the Globus Toolkit Version 3 Page 2 Colorado Software Summit: October 26 – 31, 2003 © Copyright 2003, IBM Corporation Introduction to Grid Computing Don't panic: it's just the new word for distributed computing We want to exploit resources on a wide, loose network ➢ Not a cluster, but heterogeneous systems ➢ CPU, storage, software, devices, all w/quality-of-service We started small in the early days with file copying, printing Later we moved to program-program communication We have to solve the same problems every time ➢ Comms protocol, data format, security, management Why can't we just focus on the high-level tasks? Please? ➢ Leverage existing standards ➢ Follow accepted patterns Michael Brown – Introduction to Grid Computing with the Globus Toolkit Version 3 Page 3 Colorado Software Summit: October 26 – 31,