Virtual Cluster Management for Analysis Of

Total Page:16

File Type:pdf, Size:1020Kb

Virtual Cluster Management for Analysis Of VIRTUAL CLUSTER MANAGEMENT FOR ANALYSIS OF GEOGRAPHICALLY DISTRIBUTED AND IMMOVABLE DATA Yuan Luo Submitted to the faculty of the University Graduate School in partial fulfillment of the requirements for the degree Doctor of Philosophy in the School of Informatics and Computing, Indiana University August 2015 Accepted by the Graduate Faculty, Indiana University, in partial fulfillment of the requirements for the degree of Doctor of Philosophy. Doctoral Committee Beth Plale, Ph.D. (Chairperson) Geoffrey Fox, Ph.D. Judy Qiu, Ph.D. Yuqing Wu, Ph.D. Philip Papadopoulos, Ph.D. August 7th, 2015 ii Copyright c 2015 Yuan Luo iii In loving memory of my grandfather. To my parents, my cousin, and all family members. iv Acknowledgements It has been a great pleasure working with the faculty, staff, and students at Indiana Univer- sity, during my seven-year PhD study. I would like to express my sincere gratitude to my advisor, Professor Beth Plale. This dissertation would never have been possible without her guidance, encouragement, challenges, and support throughout my research at Indiana Uni- versity. I am grateful for the many opportunities that Beth has provided for me to research, collaborate, publish, teach, and travel independently. I thank my other research committee members, Professor Geoffrey Fox, Professor Judy Qiu, Professor Yuqing Wu, and Dr. Philip Papadopoulos, for their invaluable discussion, ideas, and feedback. I also would like to thank Professor Xiaohui Wei, Dr. Peter Arzberger, and Dr. Wilfred Li for their mentorship before I joined Indiana University. I would never have started my PhD without their influence. I thank all current and past members in the Data to Insight Center at Indiana University for making the lab an enjoyable and stimulating place to work. I would like to name these individuals in particular: Peng Chen, Yiming Sun, Dr. You-Wei Cheah, Zong Peng, Jiaan Zeng, Guangchen Ruan, Quan Zhou, Dr. Scott Jensen, Dr. Devarshi Ghoshal, Felix Terkhorn, Jenny Olmes-Stevens, Jodi Stern, and Robert Ping. I also thank Tak-Lon Wu and Zhenhua Guo from Digital Science Center for multiple discussions. v I want to thank all mentors, colleagues and friends in the PRAGMA community for research discussions, project collaborations, and hospitalities: Shava Smallen, Dr. Kevin Kejun Dong, Professor Jose Fortes, Nadya Williams, Cindy Zheng, Dr. Yoshio tanaka, Professor Osamu Tatebe, Teri Simas, and many more. I thank all faculty and staff at School of Informatics and Computing for the excellent teaching and service. I am grateful to the school for providing me academic fellowship at the time of enrollment. I would like to thank Jenny Olmes-Stevens, Marisha Zimmerman, and Yi Qiao for all their help in thesis proofreading. Finally, I thank my family for all of the love, support, encouragement, and prayers throughout my journey. vi Yuan Luo VIRTUAL CLUSTER MANAGEMENT FOR ANALYSIS OF GEOGRAPHICALLY DISTRIBUTED AND IMMOVABLE DATA Scenarios exist in the era of Big Data where computational analysis needs to utilize widely distributed and remote compute clusters, especially when the data sources are sensitive or extremely large, and thus unable to move. A large dataset in Malaysia could be ecologically sensitive, for instance, and unable to be moved outside the country boundaries. Controlling an analysis experiment in this virtual cluster setting can be difficult on multiple levels: with setup and control, with managing behavior of the virtual cluster, and with interoperability issues across the compute clusters. Further, datasets can be distributed among clusters, or even across data centers, so that it becomes critical to utilize data locality information to optimize the performance of data-intensive jobs. Finally, datasets are increasingly sensitive and tied to certain administrative boundaries, though once the data has been processed, the aggregated or statistical result can be shared across the boundaries. This dissertation addresses management and control of a widely distributed virtual cluster having sensitive or otherwise immovable data sets through a controller. The Virtual Cluster Controller (VCC) gives control back to the researcher. It creates virtual clusters across multiple cloud platforms. In recognition of sensitive data, it can establish a single network overlay over widely distributed clusters. We define a novel class of data, notably immovable data that we call “pinned data”, where the data is treated as a first-class citizen vii instead of being moved to where needed. We draw from our earlier work with a hierarchical data processing model, Hierarchical MapReduce (HMR), to process geographically distributed data, some of which are pinned data. The applications implemented in HMR use extended MapReduce model where computations are expressed as three functions: Map, Reduce, and GlobalReduce. Further, by facilitating information sharing among resources, applications, and data, the overall performance is improved. Experimental results show that the overhead of VCC is minimum. The HMR outperforms traditional MapReduce model while processing a particular class of applications. The evaluations also show that information sharing between resources and application through the VCC shortens the hierarchical data processing time, as well satisfying the constraints on the pinned data. Beth Plale, Ph.D. (Chairperson) Geoffrey Fox, Ph.D. Judy Qiu, Ph.D. Yuqing Wu, Ph.D. Philip Papadopoulos, Ph.D. viii Contents 1 Introduction 1 1.1 Motivation . .1 1.2 Challenges and Contributions . .5 1.2.1 Multi-Cloud Controllability and Interoperability . .6 1.2.2 Manage, Access, and Process of Pinned Data . .8 1.3 Definition of Terminologies . 10 1.4 Dissertation Outline . 11 2 Related Work 12 2.1 Multi-Cloud Controllability and Interoperability . 12 2.2 Manage, Access, and Use of Pinned Data . 16 2.3 Background Related Work . 19 3 Background Work 24 3.1 Hierarchical MapReduce Framework . 24 3.1.1 Architecture . 25 3.1.2 Programming Model . 26 3.2 HMR Job Scheduling . 29 3.2.1 Compute Capacity Aware Scheduling . 30 ix 3.2.2 Data Location Aware Scheduling . 32 3.2.3 Topology and Data Location Aware Scheduling . 35 3.3 Hierarchical MapReduce Applications . 37 3.3.1 Compute-Intensive Applications . 37 3.3.2 Data-Intensive Applications . 39 3.4 HMR Framework Evaluation . 41 3.4.1 Evaluation of AutoDock HMR . 41 3.4.2 Gfarm Over Hadoop Clusters . 47 4 Virtual Cluster Controller 49 4.1 Architecture . 50 4.1.1 VCC Design . 51 4.1.2 VCC Implementation . 51 4.1.3 PRAGMA Cloud . 55 4.2 Specifications of Cloud Resources, Applications and Data . 55 4.2.1 Deployment vs Execution . 56 4.2.2 Resource, Application and Data Specifications . 57 4.2.3 Runtime Information for Provisioned Resource . 58 4.3 VCC Resource Allocation: A Matchmaking Process . 62 4.3.1 Application Aware Resource Allocation . 62 4.3.2 VCC Matchmaker . 63 4.4 VCC Framework Evaluation . 68 4.4.1 Testbed . 69 4.4.2 VCC Overhead Evaluation . 70 x 4.4.3 Network Overhead Evaluation . 71 4.5 Information Sharing Evaluation . 75 4.6 Summary . 78 5 Manage, Access, and Use of Pinned Data 82 5.1 Introduction . 83 5.2 Pinned Data Model . 83 5.2.1 Pinned Data (PND) . 83 5.2.2 Pinned Data Suitcase (PDS) . 84 5.2.3 Access Pinned Data . 91 5.3 Pinned Data Processing with HMR+ . 91 5.3.1 HMR+ Architecture . 91 5.3.2 Pinned-Data Applications . 94 5.4 Pinned Data Evaluation . 94 5.4.1 Experimental Setup . 95 5.4.2 PDS Creation Evaluation . 96 5.4.3 PDS Server Evaluation . 98 5.4.4 Pinned Data on HMR+ Evaluation . 102 5.5 Summary . 106 6 Conclusion and Future Work 110 6.1 Summary . 110 6.2 Conclusion . 111 6.3 Future Work . 112 xi Bibliography 113 Curriculum Vitae xii List of Figures 3.1 Hierarchical MapReduce Architecture. 27 3.2 HMR Data Flow. 28 3.3 HMR Execution Steps. 41 3.4 Local cluster MapReduce execution time based on different number of Map tasks. 43 3.5 (a) Two-way data movement cost of γ-weighted partitioned datasets: local MapReduce inputs and outputs; (b) Local MapReduce turnaround time of γ-weighted datasets, including data movement cost. 44 3.6 (a) Two-way data movement cost of γθ-weighted partitioned datasets: local MapReduce inputs and outputs; (b) Local MapReduce turnaround time of γθ-weighted datasets, including data movement cost. 45 4.1 Resource management system for hybrid cloud infrastructures. 50 4.2 Virtual Cluster Controller Architecture . 52 4.3 Architecture of VCC-HTCondor Pool . 53 4.4 VCC Enabled PRAGMA Cloud . 54 4.5 Runtime topology information of a virtual cluster. VCk is a subset of a virtual cluster on the physical cluster k.Xij represents network bandwidth (Bij) and latency (Lij) measured from VCi to VCj.............. 61 xiii 4.6 Resource Allocation Model . 62 4.7 Matchmaker Architecture . 66 4.8 Matchmaker Engine flowchart . 67 4.9 VCC overhead with ad-hoc IPOP installation . 72 4.10 VCC overhead with IPOP pre-installed . 72 4.11 TCP/UDP Bandwidth and RTT Tests . 73 4.12 TCP/UDP Bandwidth and RTT Tests over IPOP . 73 4.13 Data-Intensive HMR application execution time using 4 algorithms combi- nations under different data distribution. The data transfer speed between IU and SDSC is 3MB/s. 79 4.14 Data-Intensive HMR application execution time using 4 algorithms combi- nations under different data distribution. The data transfer speed between IU and SDSC is 5MB/s. 79 4.15 Data-Intensive HMR application execution time using 4 algorithms combi- nations under different data distribution. The data transfer speed between IU and SDSC is 7MB/s. 80 4.16 Speedup of AARA+TDLAS algorithms combination over other algorithms combinations. The x-axis indicates the data transfer speed ratio of LAN overWAN. .................................. 80 5.1 Pinned Data Suitcase with Local Data .
Recommended publications
  • Push-Based Job Submission Using Reverse SSH Connections
    rvGAHP – Push-Based Job Submission using Reverse SSH Connections Scott Callaghan Gideon Juve Karan Vahi University of Southern California USC Information Sciences Institute USC Information Sciences Institute Los Angeles, California Marina Del Rey, California Marina Del Rey, California [email protected] [email protected] [email protected] Philip J. Maechling Thomas H. Jordan Ewa Deelman University of Southern California University of Southern California USC Information Sciences Institute Los Angeles, California Los Angeles, California Marina Del Rey, California [email protected] [email protected] [email protected] ABSTRACT using Reverse SSH Connections. In WORKS’17: WORKS’17: 12th Workshop Computational science researchers running large-scale scientific on Workflows in Support of Large-Scale Science, November 12–17, 2017, Denver, CO, USA. ACM, New York, NY, USA, 8 pages. https://doi.org/10.1145/3150994. workflow applications often want to run their workflows onthe 3151003 largest available compute systems to improve time to solution. Workflow tools used in distributed, heterogeneous, high perfor- mance computing environments typically rely on either a push- 1 INTRODUCTION based or a pull-based approach for resource provisioning from these Modern scientific applications typically require the execution of compute systems. However, many large clusters have moved to multiple codes, may include computational and data dependencies, two-factor authentication for job submission, making traditional au- and contain varied computational models ranging from a bag of tomated push-based job submission impossible. On the other hand, tasks to a monolithic parallel code. These applications are often pull-based approaches such as pilot jobs may lead to increased com- complex software suites with substantial computational and data plexity and a reduction in node-hour efficiency.
    [Show full text]
  • Cloud Computing Bible Is a Wide-Ranging and Complete Reference
    A thorough, down-to-earth look Barrie Sosinsky Cloud Computing Barrie Sosinsky is a veteran computer book writer at cloud computing specializing in network systems, databases, design, development, The chance to lower IT costs makes cloud computing a and testing. Among his 35 technical books have been Wiley’s Networking hot topic, and it’s getting hotter all the time. If you want Bible and many others on operating a terra firma take on everything you should know about systems, Web topics, storage, and the cloud, this book is it. Starting with a clear definition of application software. He has written nearly 500 articles for computer what cloud computing is, why it is, and its pros and cons, magazines and Web sites. Cloud Cloud Computing Bible is a wide-ranging and complete reference. You’ll get thoroughly up to speed on cloud platforms, infrastructure, services and applications, security, and much more. Computing • Learn what cloud computing is and what it is not • Assess the value of cloud computing, including licensing models, ROI, and more • Understand abstraction, partitioning, virtualization, capacity planning, and various programming solutions • See how to use Google®, Amazon®, and Microsoft® Web services effectively ® ™ • Explore cloud communication methods — IM, Twitter , Google Buzz , Explore the cloud with Facebook®, and others • Discover how cloud services are changing mobile phones — and vice versa this complete guide Understand all platforms and technologies www.wiley.com/compbooks Shelving Category: Use Google, Amazon, or
    [Show full text]
  • Introduction to Python University of Oxford Department of Particle Physics
    Particle Physics Cluster Infrastructure Introduction University of Oxford Department of Particle Physics October 2019 Vipul Davda Particle Physics Linux Systems Administrator Room 661 Telephone: x73389 [email protected] Particle Physics Computing Overview 1 Particle Physics Linux Infrastructure Distributed File System gluster NFS /data Worker Nodes /data/atlas /data/lhcb NFS HTCondor /home Batch Server Interactive Servers physics_s/eduroam Network Printer Managed Laptops Managed Desktops Particle Physics Computing Overview 2 Introduction to the Unix Operating System Unix is a Multi-User/Multi-Tasking operating system. Developed in 1969 at AT&T’s Bell Labs by Ken Thompson (Unix) Dennis Ritchie (C) Unix is written in C programming language. Unix was originally a command-line OS, but now has a graphical user interface. It is available in many different forms: Linux , Solaris, AIX, HP-UX, freeBSD It is a well-suited environment for program development: C, C++, Java, Fortran, Python… Unix is mainly used on large servers for scientific applications. Particle Physics Computing Overview 3 Linux Distributions Source: https://www.muylinux.com/2009/04/24/logos-de-distribuciones-gnulinux/ Particle Physics Computing Overview 4 Particle Physics Linux Infrastructure Particle Physics uses CentOS Linux on the cluster. It is a free version of RedHat Enterprise Linux. Particle Physics Computing Overview 5 Basic Linux Commands Particle Physics Computing Overview 6 Basic Linux Commands o ls - list directory contents ls –l - long listing
    [Show full text]
  • The Translational Journey of the Htcondor-CE
    Journal of Computational Science xxx (xxxx) xxx Contents lists available at ScienceDirect Journal of Computational Science journal homepage: www.elsevier.com/locate/jocs Principles, technologies, and time: The translational journey of the HTCondor-CE Brian Bockelman a,*, Miron Livny a,b, Brian Lin b, Francesco Prelz c a Morgridge Institute for Research, Madison, USA b Department of Computer Sciences, University of Wisconsin-Madison, Madison, USA c INFN Milan, Milan, Italy ARTICLE INFO ABSTRACT Keywords: Mechanisms for remote execution of computational tasks enable a distributed system to effectively utilize all Distributed high throughput computing available resources. This ability is essential to attaining the objectives of high availability, system reliability, and High throughput computing graceful degradation and directly contribute to flexibility, adaptability, and incremental growth. As part of a Translational computing national fabric of Distributed High Throughput Computing (dHTC) services, remote execution is a cornerstone of Distributed computing the Open Science Grid (OSG) Compute Federation. Most of the organizations that harness the computing capacity provided by the OSG also deploy HTCondor pools on resources acquired from the OSG. The HTCondor Compute Entrypoint (CE) facilitates the remote acquisition of resources by all organizations. The HTCondor-CE is the product of a most recent translational cycle that is part of a multidecade translational process. The process is rooted in a partnership, between members of the High Energy Physics community and computer scientists, that evolved over three decades and involved testing and evaluation with active users and production infrastructures. Through several translational cycles that involved researchers from different organizations and continents, principles, ideas, frameworks and technologies were translated into a widely adopted software artifact that isresponsible for provisioning of approximately 9 million core hours per day across 170 endpoints.
    [Show full text]
  • Manager, Software Engineering
    RESUME RAMESH A (PRINCE2® Practitioner) E-mail : [email protected] Mobile : +919886311312 Summary: Over 14 years 10 months of involvement in IT industry with solid foundation on Software Testing (as Manager, Test/Technical lead, Test Architect, Scrum Master) in the cutting edge innovations/technologies Managing, Mentoring, Guiding and Leading 14 QA team members across 4 different projects Implementing QA strategies, Open source technologies to maximize the Product Quality and Test Coverage Accountable and Responsible for planning, managing, executing the complete End to End QE activities (Starting from Requirements gathering to QE Sign-off) Open source contributor for ManageIQ, Aeolus, Deltacloud API, Open Stack Well experienced in Designing automation framework using Selenium with Java and Python Possess rich experience in Design, Development and Testing with excellent analytical, problem solving, communication and interpersonal skills. Well aware of working with both Upstream(open source community) and Downstream(Enterprise release) Techno-functional with sound knowledge in management of various activities including development/ testing/ deployment/ configurations/ maintenance of an enterprise wide Operating System, Cloud applications, Middleware application, functional testing, API testing, non-functional testing, UAT, Automation and end-user trainings Experienced in writing/ maintaining test plans, test strategies, test cases, wiki pages and docs for the functionality, installation/ configuration, automation setup and
    [Show full text]
  • Innovation Across the Open Hybrid Cloud Red Hat Summit 2018 Press Conference
    INNOVATION ACROSS THE OPEN HYBRID CLOUD RED HAT SUMMIT 2018 PRESS CONFERENCE Paul Cormier Matt Hicks President, Products and Technologies SVP, Engineering Red Hat Red Hat Ashesh Badani Mike Ferris VP and General Manager, OpenShift VP, Technical Business Development & Red Hat Business Architecture Red Hat RED HAT’S INTENTIONAL 25-YEAR JOURNEY 1993 FOUNDED 2012 $1 BILLION IN REVENUE RED HAT STORAGE RELEASED 1999 IPO FUSESOURCE, POLYMITA & MANAGEIQ ACQUIRED 2002 FIRST RELEASE OF ENTERPRISE LINUX 2013 RED HAT OPENSTACK PLATFORM RELEASED OPENSHIFT ENTERPRISE RELEASED 2006 JBOSS ACQUIRED 2014 INKTANK (CEPH), ENOVANCE (OPENSTACK), 2009 RED HAT VIRTUALIZATION RELEASED & FEEDHENRY (MOBILE) ACQUIRED RED HAT ADDED TO S&P 500 INDEX 2015 ANSIBLE ACQUIRED 2011 2016 $2 BILLION IN REVENUE GLUSTER ACQUIRED OPENSHIFT RELEASED 3SCALE (API MANAGEMENT) ACQUIRED 2017 PERMABIT & CODENVY ACQUIRED COREOS ACQUIRED 2018 $3 BILLION ANNUAL RUN RATE REVENUE RED HAT SUMMIT 2018 NEWS ● REAL ENTERPRISE ADOPTION ● NEW TECHNOLOGY INNOVATIONS TO ADVANCE THE HYBRID CLOUD ● DEVELOPER MOMENTUM ● MOMENTUM ACROSS THE CLOUD-NATIVE ISV AND HYBRID CLOUD ECOSYSTEM THE 3 PILLARS OF RED HAT’S BUSINESS SUPPORTED BY AN ENTIRE TECHNOLOGY ECOSYSTEM We have the Linux We have the leading We have the foundation & the cloud enterprise Kubernetes management & platforms to win hybrid container platform with automation solutions to cloud infrastructure middleware services to make our portfolio sticky win the developer & easier to use WE HAVE THE PARTNER ECOSYSTEM TO WIN OPEN HYBRID CLOUD RED HAT MAKES THE HYBRID CLOUD AND CONTAINER-NATIVE ENTERPRISE A REALITY RED HAT ENABLES TRANSFORMATION ACROSS INDUSTRIES ANNOUNCING: NEW TECHNOLOGY INNOVATIONS TO ADVANCE THE HYBRID CLOUD HYBRID CLOUD INFRASTRUCTURE SUMMIT NEWS & DEMOS NEW - CoreOS INTEGRATION: OPENSHIFT AND RED HAT CoreOS HYBRID CLOUD NEW - OPENSHIFT+OPENSTACK: INTEGRATING HYBRID INFRASTRUCTURE CLOUD INFRASTRUCTURE WITH CLOUD-NATIVE APP DEV Infrastructure software across the 4 footprints, with DEMO - TOOLING AND SERVICES TO MIGRATE FROM VMware RHEL at the very core.
    [Show full text]
  • Connor Penhale Enterprise Software Architect
    Connor Penhale Enterprise Software Architect mailto: [email protected] ​ Open Source & Cloud Evangelist Tel:+13035526680 (mobile) Servant Leader 6421 W 72nd Dr Arvada, CO 80003 Entrepreneur Executive Summary: ● Developing Enterprise Applications utilizing Java EE and Messaging since 2005 ● Fortune 500 Experience: Bank of America, Coca Cola, CVS, Home Depot, Wells Fargo ● Founded Startup in 2014, $150k fundraising, $150k revenue, 5 staff, 13k+ personal man hours Battle-Tested Experience and Deep Technical Acumen: ● Enterprise Integration Patterns, Distributed Computing, and Messaging with technologies like Apache Camel, JBoss / Wildfly, ElasticSearch, JMS, Websockets, REST, Nginx, PostgresSQL ● DevSecOps in the Cloud, on-premise, and in hybrid environments with technologies like OpenShift, Kubernetes, Jenkins, Oauth, Puppet, Ansible, ManageIQ, Foreman, CloudFormation ● Design, Deployment, and Operation of On-Premise Datacenters with technologies like OpenStack, OVirt, Ceph, Cinder, iSCSI, SAN, Hyper Converged Infrastructure, HA, DR Professional Experience: Rogue Wave Software - OpenLogic – Louisville, CO 10/2018-Present Enterprise Architect 12/2012-04/2015 I joined OpenLogic in December of 2012, and got to realize my dream of being an open source evangelist as a full-time aspect of my professional duties. By bringing the white-glove service I perfected at Polycom to customers like Bank of America, Coca Cola, CVS, FirstData, Home Depot, and Wells Fargo, I was able to provide an incredible value to the team, and gain exposure at an architecture level to the best running and most challenging networks of applications in the Fortune 500. I’ve returned to this exciting role to spearhead the roll-out of Rogue Wave Software’s curated Cloud Native stacks. Turnberry Solutions – Englewood, CO 02/2018-10/2018 Java Lead Embedded at Comcast to tackle ETL business requirements through event-driven programming using core Java, Spring Boot, Camel ESB, Kafka, Avro, and other technologies.
    [Show full text]
  • Red Hat Cloudforms 5.0 Provisioning Virtual Machines and Instances
    Red Hat CloudForms 5.0 Provisioning Virtual Machines and Instances Provisioning, workload management, and orchestration for Red Hat CloudForms Last Updated: 2020-08-05 Red Hat CloudForms 5.0 Provisioning Virtual Machines and Instances Provisioning, workload management, and orchestration for Red Hat CloudForms Red Hat CloudForms Documentation Team [email protected] Legal Notice Copyright © 2020 Red Hat, Inc. The text of and illustrations in this document are licensed by Red Hat under a Creative Commons Attribution–Share Alike 3.0 Unported license ("CC-BY-SA"). An explanation of CC-BY-SA is available at http://creativecommons.org/licenses/by-sa/3.0/ . In accordance with CC-BY-SA, if you distribute this document or an adaptation of it, you must provide the URL for the original version. Red Hat, as the licensor of this document, waives the right to enforce, and agrees not to assert, Section 4d of CC-BY-SA to the fullest extent permitted by applicable law. Red Hat, Red Hat Enterprise Linux, the Shadowman logo, the Red Hat logo, JBoss, OpenShift, Fedora, the Infinity logo, and RHCE are trademarks of Red Hat, Inc., registered in the United States and other countries. Linux ® is the registered trademark of Linus Torvalds in the United States and other countries. Java ® is a registered trademark of Oracle and/or its affiliates. XFS ® is a trademark of Silicon Graphics International Corp. or its subsidiaries in the United States and/or other countries. MySQL ® is a registered trademark of MySQL AB in the United States, the European Union and other countries. Node.js ® is an official trademark of Joyent.
    [Show full text]
  • Use of Open Source Cloud Technologies to Deliver Modern Public Health Services
    Use of open source cloud technologies to deliver modern public health services Francesco Giannoccaro 29th January 2020 About PHE ► Public Health England (PHE) is an executive agency of the Department of Health in the UK. PHE provide government, local government, the NHS, industry and the public with evidence-based professional, scientific and delivery expertise and support. ► Public Health England was established in 2013 to bring together public health specialists from more than 70 organisations into a single public health service. PHE employee about 5,500 staff, mostly scientists, researchers and public health professionals. ► PHE mission is to protect and improve the nation’s health and wellbeing, and reduce health inequalities. We do this through world-leading science, knowledge and intelligence, advocacy, partnerships and the delivery of specialist public health services. 2 Use of open-source technologies to deliver modern public health services Francesco Giannoccaro - London 2020/01/29 HIGHLIGHT THE AMBITIOUS AND INSPIRING MISSION PUBLIC HEALTH ENGLAND HAS. HOW THIS MISSION ALIGN TO THE OPEN SOURCE VALUES, IN THAT PHE AIMS TO DELIVER INNOVATIVE PUBLIC HEALTH SERVICES TO EVERYONE INDEPENDENTLY HOW RICH THEY ARE, REDUCING HEALTH INEQUALITY. Wide range of public health services PHE deliver a wide range of public health services including • research and scientific publications based on mathematical models such as Spatial Metapopulation Model for transmissible disease (eg Flu/Smallpox), predictive models applied to the Anthrax, inference problem to be able to infer: likely size of outbreak, location of source, spatial extent, etc • pathogen genomics service, based on whole genome sequencing, for pathogen typing, surveillance and outbreak investigation.
    [Show full text]
  • Leveraging Containers and Openstack
    Leveraging Containers and OpenStack A Comprehensive Review Introduction Imagine that you are tasked to build an entire private cloud infrastructure from the ground up. You have a limited budget, a small but dedicated team, and are asked to pull off a miracle. A few years ago, you’d build an infrastructure with applications running in virtual machines, with some bare-metal machines for legacy applications. As infrastructure has evolved, virtual machines (VMs) enabled greater levels of efficiency and agility, but VMs alone don’t completely meet the needs of an agile approach to application deployment. They continue to serve as a foundation for running many applications, but increasingly, developers are looking toward the emerging trend of containers for leading-edge application development and deployment because containers offer increased levels of agility and efficiency. Container technologies like Docker and Kubernetes are becoming the leading standards for building containerized applications. They help free organizations from complexity that limits development agility. Containers, container infrastructure, and container deployment technologies have proven themselves to be very powerful abstractions that can be applied to a number of different use cases. Using something like Kubernetes, an organization can deliver a cloud that solely uses containers for application delivery. But a leading-edge private cloud isn’t just about containers, and containers aren’t appropriate for all workloads and use cases. Today, most private cloud infrastructures need to encompass bare-metal machines for managing infrastructure, virtual machines for legacy applications, and containers for newer applications. The ability to support, manage and orchestrate all three approaches is the key to operational efficiency.
    [Show full text]
  • A Comprehensive Perspective on Pilot-Job Systems
    A Comprehensive Perspective on Pilot-Job Systems ∗ Matteo Turilli Mark Santcroos Shantenu Jha RADICAL Laboratory, ECE RADICAL Laboratory, ECE RADICAL Laboratory, ECE Rutgers University Rutgers University Rutgers University New Brunswick, NJ, USA New Brunswick, NJ, USA New Brunswick, NJ, USA [email protected] [email protected] [email protected] Pilot-Jobs provide a multi-stage mechanism to execute ABSTRACT workloads. Resources are acquired via a placeholder job and subsequently assigned to workloads. Pilot-Jobs are having Pilot-Job systems play an important role in supporting dis- a high impact on scientific and distributed computing [1]. tributed scientific computing. They are used to consume more They are used to consume more than 700 million CPU hours than 700 million CPU hours a year by the Open Science Grid a year [2] by the Open Science Grid (OSG) [3, 4] communi- communities, and by processing up to 1 million jobs a day for ties, and process up to 1 million jobs a day [5] for the ATLAS the ATLAS experiment on the Worldwide LHC Computing experiment [6] on theLarge Hadron Collider (LHC) [7] Com- Grid. With the increasing importance of task-level paral- puting Grid (WLCG) [8, 9]. A variety of Pilot-Job systems lelism in high-performance computing, Pilot-Job systems are used on distributed computing infrastructures (DCI): are also witnessing an adoption beyond traditional domains. Glidein/GlideinWMS [10, 11], the Coaster System [12], DI- Notwithstanding the growing impact on scientific research, ANE [13], DIRAC [14], PanDA [15], GWPilot [16], Nim- there is no agreement upon a definition of Pilot-Job system rod/G [17], Falkon [18], MyCluster [19] to name a few.
    [Show full text]
  • EGI Federated Platforms Supporting Accelerated Computing
    EGI federated platforms supporting accelerated computing PoS(ISGC2017)020 Paolo Andreetto Jan Astalos INFN, Sezione di Padova Institute of Informatics Slovak Academy of Sciences Via Marzolo 8, 35131 Padova, Italy Bratislava, Slovakia E-mail: [email protected] E-mail: [email protected] Miroslav Dobrucky Andrea Giachetti Institute of Informatics Slovak Academy of Sciences CERM Magnetic Resonance Center Bratislava, Slovakia CIRMMP and University of Florence, Italy E-mail: [email protected] E-mail: [email protected] David Rebatto Antonio Rosato INFN, Sezione di Milano CERM Magnetic Resonance Center Via Celoria 16, 20133 Milano, Italy CIRMMP and University of Florence, Italy E-mail: [email protected] E-mail: [email protected] Viet Tran Marco Verlato1 Institute of Informatics Slovak Academy of Sciences INFN, Sezione di Padova Bratislava, Slovakia Via Marzolo 8, 35131 Padova, Italy E-mail: [email protected] E-mail: [email protected] Lisa Zangrando INFN, Sezione di Padova Via Marzolo 8, 35131 Padova, Italy E-mail: [email protected] While accelerated computing instances providing access to NVIDIATM GPUs are already available since a couple of years in commercial public clouds like Amazon EC2, the EGI Federated Cloud has put in production its first OpenStack-based site providing GPU-equipped instances at the end of 2015. However, many EGI sites which are providing GPUs or MIC coprocessors to enable high performance processing are not directly supported yet in a federated manner by the EGI HTC and Cloud platforms. In fact, to use the accelerator cards capabilities available at resource centre level, users must directly interact with the local provider to get information about the type of resources and software libraries available, and which submission queues must be used to submit accelerated computing workloads.
    [Show full text]