Hepix Report

Total Page:16

File Type:pdf, Size:1020Kb

Hepix Report HEPiX Report Helge Meinhard, Pawel Grzywaczewski, Romain Wartel / CERN-IT Post-C5/Computing Seminar 03 December 2010 CERN IT Department CH-1211 Genève 23 Switzerland www.cern.ch/it Outline • Meeting organisation, site reports, (benchmarking,) infrastructure (Helge Meinhard) • Storage, OS and applications, miscellaneous (Pawel Grzywaczewski) • Virtualisation, security and networking, grid and cloud (Romain Wartel) HEPiX report – Helge.Meinhard at cern.ch – 03-Dec-2010 HEPiX • Global organisation of service managers and support staff providing computing facilities for HEP • Covering all platforms of interest (Unix/Linux, Windows, Grid, …) • Aim: Present recent work and future plans, share experience, advise managers • Meetings ~ 2 / y (spring in Europe, autumn typically in North America) HEPiX report – Helge.Meinhard at cern.ch – 03-Dec-2010 HEPiX Autumn 2010 (1) • Held 01 – 05 November at Cornell University, Ithaca NY – CESR: Electron-positron storage ring; CLEO: experiment doing a lot of interesting b physics – New player in the HEPiX field at as site… but a well-known face: Chuck Boeheim, the previous north-American co-chair of HEPiX – Good local organisation – Nice auditorium in conference hotel, basically unlimited coffee supply – First face-to-face meeting in 2010 for most participants HEPiX report – Helge.Meinhard at cern.ch – 03-Dec-2010 HEPiX Autumn 2010 (2) • Format: Pre-defined tracks with conveners and invited speakers per track – Still room for spontaneous talks – either fit into one of the tracks, or classified as ‘miscellaneous’ – Again proved to be the right approach; in view of the low number of participants, an extremely rich, interesting and packed agenda – Judging by number of submitted abstracts, no real hot spot: 8 infrastructure, 8 Grid/clouds, 7 storage, 6 virtualisation, 6 OS and apps, 5 network and security, 3 miscellaneous, 1 benchmarking – Some abstracts submitted late, planning difficult • Full details and slides: http://indico.cern.ch/conferenceDisplay.py?confId=92498 • Trip report by Alan Silverman available, too http://cdsweb.cern.ch/record/1307061 HEPiX report – Helge.Meinhard at cern.ch – 03-Dec-2010 HEPiX Autumn 2010 (3) • 47 registered participants, of which 11 from CERN – Barring, Bell, Grzywaczewski, Janyst, Kelemen, Meinhard, Salter, Schwickerath, Silverman, T Smith, Wartel – Other sites: ASGC, Caspur, CEA, CNAF, Cornell, DESY Hamburg, DESY Zeuthen, FNAL, FZU, IN2P3, INFN Milano, INFN Pavia, JLAB, KISTI, LAL, NIKHEF, RAL, SFU, SLAC, TRIUMF, Umea U – Compare with Berkeley (autumn 2009): 61 participants, of which 9 from CERN HEPiX report – Helge.Meinhard at cern.ch – 03-Dec-2010 HEPiX Autumn 2010 (4) • 62 talks, of which 19 from CERN – Compare with Berkeley: 62 talks, of which 16 from CERN • Next meetings: – Spring 2011: GSI Darmstadt (May 2 - 6) • Possibly followed by an LCG workshop over the weekend – Autumn 2011: Vancouver (to be confirmed, date to be decided; 20th anniversary of HEPiX!) HEPiX report – Helge.Meinhard at cern.ch – 03-Dec-2010 Oracle/SUN Policy Concerns • Recent observations: – Significantly increased HW prices for Thor-style machines – Very significantly increased maintenance fees for Oracle (ex-Sun) software running on non-Oracle hardware • Sun GridEngine, Lustre, OpenSolaris, Java, OpenOffice, VirtualBox, … – Very limited collaboration with non-Oracle developers – Most Oracle software has already got forked as open- source projects – (At least) two Oracle-independent consortia around Lustre • HEP labs very concerned HEPiX report – Helge.Meinhard at cern.ch – 03-Dec-2010 Site Reports (1) • Worker node acquisitions: HPC-style small form-factor boxes (e.g. 4 dual CPU systems in a 2U enclosure) very popular – HP, Dell, Supermicro, Acer, … – Overheating CPUs because of missing thermal grease… • Disk storage – Many sites using storage-in-a-box (a la CERN or Thumper/Thor) – Some dedicated storage with SAN (FC or iSCSI) or NAS uplink • Dell MD1000, DDN 6620 (60 drives in 4U) – ASGC reporting servers of 160 TB with 10GE HEPiX report – Helge.Meinhard at cern.ch – 03-Dec-2010 Site Reports (2) • Tape storage: Large sites appear to prefer SL8500 robots and LTO5 drives • Networking: OPN hasn’t reached out yet to Norwegian and Slowenian parts of Nordic T1 • GPUs mentioned only for non-HEP applications HEPiX report – Helge.Meinhard at cern.ch – 03-Dec-2010 Site Reports (3) • Batch schedulers: One singularity (BQS) gets eradicated at last… – Replaced by (Sun|Oracle) Grid Engine • Configuration tools: Random walk of Quattor, cfengine, Puppets, … – One more very positive Quattor report from RAL • Monitoring: Many sites using Nagios, appears to become a de-facto standard • Drupal, Jabber, Trac mentioned a number of times HEPiX report – Helge.Meinhard at cern.ch – 03-Dec-2010 Site Reports (4) • Other interesting points – DESY: turning into a centre for accelerator research, particle physics and photon physics. New requirements • MacOS, Windows HPC • Large RAM machines (~500 GB!) • Lustre on MacOS and Windows • 20…100 GB from XFEL from 2012/13 on • NUMA, GPU computing • Hundreds of VOs with a handful users each – SLAC: Similar conversion as DESY • Mixed experience with back-charging for scientific computing • Tender with fixed budget for maximal capacity HEPiX report – Helge.Meinhard at cern.ch – 03-Dec-2010 Site Reports (5) –FNAL • Return air is 23 deg C, too low • DOE review recommends physics capacity without UPS – JLAB: Using tool (Surveyer) to control power consumption of desktop PCs – large savings – RAL: Quibble about email addresses – GSI: Compute requirements similar as for LHC • “Cube” computer centre prototyped – PUE 1.1 – FZU Prag: Experience with a 100 A circuit- breaker tripping… • … because it has been running at 96 A constantly… HEPiX report – Helge.Meinhard at cern.ch – 03-Dec-2010 Infrastructure (1) • 8 talks, 3 from CERN – LSF scalability tests (Schwickerath) – CERN-IT procurement (Barring) – CERN computer centre upgrade project (Salter) • Update on Quattor at RAL (Collier) – Started with new batch worker nodes. Batch done, now covering disk servers gradually – Good experience, estimate saving 0.3…0.5 FTE – Good experience with QWG as well – NIKHEF starting tests, too HEPiX report – Helge.Meinhard at cern.ch – 03-Dec-2010 Infrastructure (2) • Assets server at LEPP, Cornell U (Pulver) – OCSNG: multi-platform “moment in time” hardware and software inventory – GLPI: full lifecycle asset management – ZENOSS: agent-less monitoring, extensible via Python scripts • Scientific computing at JLAB (Philpott) – Extensive experience with GPUs: mixture of gaming cards and professional units • Mostly using CUDA, interested in OpenCL – Lustre: 300 TB on commodity HW – lots of fun points HEPiX report – Helge.Meinhard at cern.ch – 03-Dec-2010 HEPiX report – Helge.Meinhard at cern.ch – 03-Dec-2010 Infrastructure (3) • Batch infrastructure resource at DESY: BIRD (Finnern) – Collecting smallish resources, parasitic batch usage based on SGE fair-share – Contributing projects are granted fair-share points • Infrastructure improvements at IN2P3 (Olivero) – Current room: new transformers, diesel, cooling unit – New building: 2 levels of 850 m2, no offices • No raised floor, no false ceiling. All services from above • Minimal electrical redundance, will perhaps use EDF offering of double dedicated connection • Construction started in April 2010, scheduled to finish by February 2011. first production in March 2011 HEPiX report – Helge.Meinhard at cern.ch – 03-Dec-2010.
Recommended publications
  • Solaris-Cluster-Businesscontinuity-168285.Pdf
    An Oracle White Paper September 2010 Oracle Solaris and Oracle Solaris Cluster: Extending Oracle Solaris for Business Continuity Oracle White Paper— Oracle Solaris and Oracle Solaris Cluster: Extending Oracle Solaris for Business Continuity Executive Summary.............................................................................1 Introduction..........................................................................................3 Traditional Solution Components ....................................................5 Oracle Solaris Cluster for Business Continuity....................................6 Oracle Solaris Cluster......................................................................6 Introduction—Basic Clustering ........................................................8 Data Availability .............................................................................10 Applications Availability .................................................................11 Network Availability .......................................................................12 Virtualization and Oracle Solaris Cluster .......................................13 Disaster Recovery—Campus and Beyond ....................................16 A Single HA and DR Solution for Multitier Oracle Applications and Databases ..........................................................17 Oracle Solaris ....................................................................................19 Resource Management .................................................................19
    [Show full text]
  • Beginner's Guide to Oracle Grid Engine 6.2 Oracle White Paper—Beginner's Guide to Oracle Grid Engine 6.2
    An Oracle White Paper August 2010 Beginner's Guide to Oracle Grid Engine 6.2 Oracle White Paper—Beginner's Guide to Oracle Grid Engine 6.2 Executive Overview ..................................................................................... 1 Introduction .................................................................................................. 1 Chapter 1: Introduction to Oracle Grid Engine ............................................ 3 Oracle Grid Engine Jobs ......................................................................... 3 Oracle Grid Engine Component Architecture .......................................... 3 Oracle Grid Engine Basics ...................................................................... 5 Chapter 2: Oracle Grid Engine Scheduler ................................................... 10 Job Selection ........................................................................................... 10 Job Scheduling ........................................................................................ 17 Other Scheduling Features ...................................................................... 18 Additional Information on Job Scheduling ............................................... 20 Chapter 3: Planning an Oracle Grid Engine Installation .............................. 21 Installation Layout .................................................................................... 21 QMaster Data Spooling ........................................................................... 22 Execution Daemon Data
    [Show full text]
  • 7.1 Task Computing 7.2 Task-Based Application Models 7
    15CS565 CLOUD COMPUTING MODULE – III High-Throughput Computing – Task Programming (Chapter 7) Task computing is a wide area of distributed system programming encompassing several different models of architecting distributed applications, A task represents a program, which require input files and produce output files as a result of its execution. Applications are then constituted of a collection of tasks. These are submitted for execution and their output data are collected at the end of their execution. This chapter characterizes the abstraction of a task and provides a brief overview of the distrib- uted application models that are based on the task abstraction. The Aneka Task Programming Model is taken as a reference implementation to illustrate the execution of bag-of-tasks (BoT) applications on a distributed infrastructure. 7.1 Task computing 7.1.1 Characterizing a task 7.1.2 Computing categories 1 High-performance computing 2 High-throughput computing 3 Many-task computing 7.1.3 Frameworks for task computing 1. Condor 2. Globus Toolkit 3. Sun Grid Engine (SGE) 4. BOINC 5. Nimrod/G 7.2 Task-based application models 7.2.1 Embarrassingly parallel applications 7.2.2 Parameter sweep applications 7.2.3 MPI applications 7.2.4 Workflow applications with task dependencies 1 What is a workflow? 2 Workflow technologies 1. Kepler, 2. DAGMan, 3. Cloudbus Workflow Management System, and 4. Offspring. 7.3 Aneka task-based programming 7.3.1 Task programming model 7.3.2 Developing applications with the task model 7.3.3 Developing a parameter sweep application 7.3.4 Managing workflows 7.1 Task computing A task identifies one or more operations that produce a distinct output and that can be isolated as a single logical unit.
    [Show full text]
  • Development of Technological Projects with the Active Participation of Women in El Salvador
    Development of technological projects with the active participation of women in El Salvador. (April 2019) Marlene Navarro Gisela Espinoza Facultad de Ciencia y Tecnología Facultad de Ciencia y Tecnología Universidad Gerardo Barrios - UGB Universidad Gerardo Barrios - UGB San Miguel, El Salvador San Miguel, El Salvador [email protected] [email protected] Resumen—El estudio tiene como objetivo general conocer las women and men working in companies participating under the diferentes soluciones tecnológicas desarrolladas con una fuerte research. participación de mujeres; a la vez conocer una breve reseña histórica Thanks to the contribution of the female gender, foundations sobre la contribución de algunas mujeres a las Tecnologías de la of many technologies were made; that nowadays are used with Información y la Comunicación (TIC), ya que no han sido great ease and allow social diversity to develop work, professional reconocidas en su momento, si no con el paso de los años y a través and personal activities. de los resultados de diferentes investigaciones que se difunden a nivel Finally, El Salvador is part of the digital transformation in mundial por la equidad de género en todas sus áreas. different sectors: education, government, business, among others, El estudio también apuntó a visibilizar a las mujeres latinas en where Salvadoran women have active and essential participation. su rol como desarrolladoras de soluciones tecnológicas a nivel académico y empresarial; también conocer casos de éxito de mujeres Index Terms— (El Salvador,
    [Show full text]
  • HP-UX to Oracle Solaris Porting Guide Getting Started on the Move to Oracle Solaris
    An Oracle White Paper October 2011 HP-UX to Oracle Solaris Porting Guide Getting Started on the Move to Oracle Solaris HP-UX to Oracle Solaris Porting Guide Chapter 1 Introduction...................................................................... 1 Oracle Solaris ................................................................................. 1 The Advantages of Porting to Oracle Solaris.................................. 2 Chapter 2 The Porting Process........................................................ 4 Infrastructure and Application Porting Assessment ........................ 4 Build Environment Deployment....................................................... 5 Data Integration .............................................................................. 5 Source Code Porting....................................................................... 6 Application Verification.................................................................... 6 Commercial Applications and Third-Party Products ....................... 6 Chapter 3 Operating System Considerations................................... 7 Processor Endianness .................................................................... 7 Data Alignment ............................................................................... 7 Read/Write Structures..................................................................... 8 Storage Order and Alignment ......................................................... 8 64-Bit Data Models ........................................................................
    [Show full text]
  • Prosabladet Udgivelse: 2011/6
    bladet ProsaDe it-professionelles fagblad · 6 · Juni · 2011 Tema: Sådan tackler vi fremtidens pressede arbejdsmarked s. 18-25 Fri fugl: Lav et fællesskab for faktura-arbejdere s. 22 1. maj i konfliktens s. 8 tegn Kom med i Fail Club s. 14 Open source-genbrug kan spare millioner s. 26 Synspunkt Prosabladet · 6 · 2011 41. årgang CSC – er etikken Adresse: Vester Farimagsgade 37A, 1606 Kbh. V. Tlf. 33 36 41 41, fax 33 91 90 44 Prosabladets postgiro: 6 58 08 90, i højsædet? e-mail: [email protected] Hanne Lykke Jespersen, næstformand, [email protected] Redaktion: Ansvarshavende redaktør Kurt Westh Nielsen [email protected] PROSA tog i starten af maj det usædvanlige skridt at melde en af vores arbejdsgivere Nina Ferdinand, journalist [email protected] – nemlig CSC - til politiet. CSC har udråbt dette som et kampskridt over for CSC som et Palle Skramsø, layout led i den verserende konflikt på arbejdspladsen. Udgivelsesdato og deadline: Udkommer en gang hver måned, undtagen august måned. Men det er helt forkert opfattet. Der er helt andre bevæggrunde for, at vi er skredet til Næste gang 30. juni så drastisk en handling. Læserbreve: 3. juni Debatindlæg og klageadgang: Sagen startede med, at en gruppe medarbejdere nedlagde arbejdet som protest mod Prosabladet modtager meget gerne læserbreve, deba- tindlæg, tips, artikelforslag og kommentarer fra med- det psykiske arbejdsmiljø, de blev budt. CSC krævede blandt andet, at vores medlem- lemmerne. Indlæggene sendes til redaktion@prosa. mer skulle oplære de folk, som senere under konflikten skulle agere som skruebræk- dk. Indlæggene vil, såfremt de offentliggøres, blive redigeret efter retningslinjer vedtaget af redaktionen.
    [Show full text]
  • Study of Scheduling Optimization Through the Batch Job Logs Analysis 윤 준 원1 · 송 의 성2* 1한국과학기술정보연구원 슈퍼컴퓨팅본부 2부산교육대학교 컴퓨터교육과
    디지털콘텐츠학회논문지 Journal of Digital Contents Society JDCS Vol. 18, No. 7, pp. 1411-1418, Nov. 2017 배치 작업 로그 분석을 통한 스케줄링 최적화 연구 Study of Scheduling Optimization through the Batch Job Logs Analysis 1 2* 윤 준 원 · 송 의 성 1한국과학기술정보연구원 슈퍼컴퓨팅본부 2부산교육대학교 컴퓨터교육과 JunWeon Yoon1 · Ui-Sung Song2* 1Department of Supercomputing Center, KISTI, Daejeon 34141, Korea 2*Department of Computer Education, Busan National University of Education, Busan 47503, Korea [요 약] 배치 작업 스케줄러는 클러스터 환경에서 구성된 계산 자원을 인지하고 순서에 맞게 효율적으로 작업을 배치하는 역할을 수행 한다. 클러스터내의 한정된 가용자원을 효율적으로 사용하기 위해서는 사용자 작업의 특성을 분석하여 반영하여야 하는데 이를 위해서는 다양한 스케줄링 알고리즘을 파악하고 해당 시스템 환경에 맞게 적용하는 것이 중요하다. 대부분의 스케줄러 소프트웨 어는 전체 관리 대상의 자원 명세와 시스템의 상태뿐만 아니라 작업 제출부터 종료까지 다양한 사용자의 작업 수행 환경을 반영하 게 된다. 또한 작업 수행과 관련한 다양한 정보 가령, 작업 스크립트, 환경변수, 라이브러리, 작업의 대기, 시작, 종료 시간 등을 저 장하게 된다. 본 연구에서는 배치 스케줄러를 통한 작업 수행과 관련된 정보를 통해 사용자의 작업 성공률, 수행시간, 자원 규모 등의 스케줄러의 수행 로그를 분석하여 문제점을 파악하였다. 향후 이 연구를 바탕으로 자원의 활용률을 높임으로써 시스템을 최 적화할 수 있다. [Abstract] The batch job scheduler recognizes the computational resources configured in the cluster environment and plays a role of efficiently arranging the jobs in order. In order to efficiently use the limited available resources in the cluster, it is important to analyze and characterize the characteristics of user tasks. To do this, it is important to identify various scheduling algorithms and apply them to the system environment.
    [Show full text]
  • Linux Control Groups Univa Grid Engine
    Linux Control Groups Support for Univa Grid Engine Daniel Gruber [email protected] © Copyright Univa 2009-2014 Proprietary and Confidential … a Little Bit of History Sun Grid Engine © Copyright Univa 2009-2014 Proprietary and Confidential … a Little Bit of History Sun Grid Engine Oracle acquired Sun © Copyright Univa 2009-2014 Proprietary and Confidential … a Little Bit of History Sun Grid Engine Oracle acquired Sun Oracle Grid Engine © Copyright Univa 2009-2014 Proprietary and Confidential … a Little Bit of History Sun Grid Engine Oracle acquired Sun Oracle Grid Engine Developers moved to Univa © Copyright Univa 2009-2014 Proprietary and Confidential … a Little Bit of History Sun Grid Engine Oracle acquired Sun Oracle Grid Engine > 3 years ACTIVE Developers moved to Univa Development Univa Grid Engine © Copyright Univa 2009-2014 Proprietary and Confidential … a Little Bit of History Sun Grid Engine Oracle acquired Sun Oracle Grid Engine > 3 years ACTIVE Developers moved to Univa Development Univa Grid Engine 2013: Univa got all source code / IP / customer support for SGE / OGE from Oracle © Copyright Univa 2009-2014 Proprietary and Confidential Many Features added ... ShareTree on Job Array Tasks Cloud Bursting Job Classes Configure PE Sorting Order PostgreSQL Spooling License Orchestrator Support Better Job Runtime Limit Control JSV Enhancements Web GUI For Accounting: UniSight included NUMA Aware Scheduling Better MPI Support RSMAP Easier Debugging Faster JSV Job Submission Time Active Roadmap! World Wide Support Improved Scalability Cray XC-30 Support ...and many more HP CMU Support GPU Support Completely New User Guide MacOS X Support Cache Size Reporting Per Socket Memory Management Memory Affinity Settings Intel Xeon Phi Support Hundreds of Bug Fixes Completely New Admin Guide Completely New Installation Guide © Copyright Univa 2009-2014 Proprietary and Confidential Many Features added ..
    [Show full text]
  • Workload Schedulers - Genesis, Algorithms and Comparisons
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE International Journal of Computer Science and Software Engineering (IJCSSE), Volume 4, Issue 6, June 2015 provided by WestminsterResearch ISSN (Online): 2409-4285 www.IJCSSE.org Page: 141-155 Workload Schedulers - Genesis, Algorithms and Comparisons Leszek Sliwko1 and Vladimir Getov2 1, 2 Faculty of Science and Technology, University of Westminster, 115 New Cavendish Street, United Kingdom [email protected], [email protected] ABSTRACT especially for online request scheduling [60]. However, In this article we provide brief descriptions of three classes of not taking into account additional factors such as current schedulers: Operating Systems Process Schedulers, Cluster server load, network infrastructure or storage availability Systems Jobs Schedulers and Big Data Schedulers. We describe may result in an inefficient utilization of available their evolution from early adoptions to modern implementations, machines and overall higher operational system costs. considering both the use and features of algorithms. In summary, More complex algorithms rely on the availability of static we discuss differences between all presented classes of schedulers and discuss their chronological development. In infrastructure data such as CPU speed, installed memory conclusion we highlight similarities in the focus of scheduling etc. As an example we find largest remaining processing strategies design, applicable to both local and distributed systems. time on fastest machine rule (LRPT-FM), where the job with the most remaining processing time is assigned to Keywords: Schedulers, Workload, Cluster, Cloud, Process, Big fastest machine [45]. LRPT_FM approach offers clear Data. advantages in heterogeneous environments, where the system is composed from machines with different configurations.
    [Show full text]
  • Univa Grid Engine Status at CC-IN2P3 Vanessa HAMAR for the CCIN2P3 Batch Team Overview
    Centre de Calcul de l’Institut National de Physique Nucléaire et de Physique des Particules 2014/03/11 Univa Grid Engine status at CC-IN2P3 Vanessa HAMAR for the CCIN2P3 Batch Team Overview } Team members } History } Configuration } Numbers } Grid integration } Monitoring } Multicore } What we want? } Future Plans 2014/03/11 2 Batch team members } Batch team leader: ◦ Suzanne Poulat } System Administrators: 1 FTE ◦ Aurélien Gounon ◦ Mattieu Puel ◦ Vanessa Hamar } User Support and UGE administrators: ◦ Bernard Chambon ◦ Nadia Lajili ◦ Rachid Lemrani 2014/03/11 3 History BQS 1992-2012 OGE 2012-2013 UGE 2013 - ... 2014/03/11 4 History BQS 1992-2012 OGE 2012-2013 UGE 2013 - ... } BQS – Home made batch system ◦ 1992 – 2012 -> 20 years of service! ◦ Stable and robust product: 10.000 simultaneous jobs 100.000 jobs per day 2010 ◦ Jobs and configuration in a MySql DB “HEP world” faced new requirements: • VM • multicores • interactive • increase of the needs 2014/03/11 5 History BQS 1992-2012 OGE 2012-2013 UGE 2013 - ... } OGE ◦ 2010 : Sun Grid Engine (SGE) chosen ◦ Used in a test farm to customize the configuration ◦ March 2011 : Farm opened to users ◦ At the beginning of 2011 : Sun bought by Oracle ◦ since July 2011 : Oracle Grid Engine (OGE) Oracle support was not satisfactory. 2014/03/11 6 History BQS 1992-2012 OGE 2012-2013 UGE 2013 - ... } UGE ◦ Fall 2012 -> Evaluation other Grid Engine options ◦ Early 2013 -> choice of Univa Grid Engine (UGE) ◦ June 2013 -> UGE replaces OGE (2013/06/05) 2014/03/11 7 Configuration - Nodes } Common ◦ Operating
    [Show full text]
  • Oracle Grid Engine: an Overview
    An Oracle White Paper August 2010 Oracle Grid Engine: An Overview Oracle White Paper—Oracle Grid Engine: An Overview Executive Overview Oracle Grid Engine is a powerful workload management tool for maximizing the business value of an organizations computing resources. By using Oracle Grid Engine, businesses can deliver their products faster, more efficiently, and with lower overall costs. This paper Introduces the Oracle Grid Engine product and some common and uncommon use cases. 1 Oracle White Paper—Oracle Grid Engine: An Overview Introduction to Workload Management A computer is a great tool for doing work. For a desktop user that work looks like editing spreadsheets, visiting web sites, and playing Tetris when no one else is looking. The job of the computer is mostly just moving letters and numbers and windows around on the screen. For a server, however, the work looks very different. A server spends its days doing things like hosting web sites, processing email messages, and running calculations. Web sites and email are known as services. They're only useful if they're always (or almost always) there. If they move around or aren't always available, no one will use them. Services are most often run on dedicated machines. Servers tend to be used for one of two purposes: running services or processing workloads. Services tend to be long-running and don't tend to move around much. Workloads, however, such as running calculations, are usually done in a more "on demand" fashion. When a user needs something, he tells the server, and the server does it.
    [Show full text]
  • Electricaccelerator® Installation and Configuration Guide, Version
    ElectricAccelerator Installation and Configuration Guide Version 9.1 Electric Cloud, Inc. 35 South Market Street, Suite 100 San Jose, CA 95113 www.electric-cloud.com ElectricAccelerator Version 9.1 Copyright © 2002–2017 Electric Cloud, Inc. All rights reserved. Published 7/24/2017 Electric Cloud® believes the information in this publication is accurate as of its publication date. The information is subject to change without notice and does not represent a commitment from the vendor. THE INFORMATION IN THIS PUBLICATION IS PROVIDED “AS IS.” ELECTRIC CLOUD, INCORPORATED MAKES NO REPRESENTATIONS OR WARRANTIES OF ANY KIND WITH RESPECT TO THE INFORMATION IN THIS PUBLICATION AND SPECIFICALLY DISCLAIMS IMPLIED WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Use, copying, and distribution of any ELECTRIC CLOUD software described in this publication requires an applicable software license. Copyright protection includes all forms and matters of copyrightable material and information now allowed by statutory or judicial law or hereinafter granted, including without limitation, material generated from software programs displayed on the screen such as icons and screen display appearance. The software and/or databases described in this document are furnished under a license agreement or nondisclosure agreement. The software and/or databases may be used or copied only in accordance with terms of the agreement. It is against the law to copy the software on any medium except as specifically allowed in the license or nondisclosure agreement.
    [Show full text]