Evaluation and Benchmarking of Singularity MPI Containers on EU Research E-Infrastructures

Total Page:16

File Type:pdf, Size:1020Kb

Evaluation and Benchmarking of Singularity MPI Containers on EU Research E-Infrastructures Evaluation and Benchmarking of Singularity MPI Containers on EU Research e-Infrastructures Víctor Sande Veiga, Manuel Simon, Abdulrahman Azab, Giuseppa Muscianisi, Carlos Fernandez, Giuseppe Fiameni, and Simone Marocchi CIMNE - International Centre for Numerical Methods in Engineering, Spain Nokia, France CESGA - Centro de Supercomputación de Galicia, Spain University of Oslo, Norway CINECA - Interuniversity Consortium, Italy 1 CANOPIE-HPC@SC19 www.mso4sc.eu | www.prace-ri.eu Partnership for Advanced Mathematical Modelling, Simulation and Computing in Europe Optimization for Societal Challenges with Scientific Computing 2 CANOPIE-HPC@SC19 www.prace-ri.eu (CESGA) Supercomputing Center of Galicia ▶ 328 TFLOPS ▶ 44,8 TB of RAM ▶ 1,5 PB of disk capacity 3 CANOPIE-HPC@SC19 www.mso4sc.eu | www.prace-ri.eu (CINECA) Interuniversity Consortium, Italy MARCONI is the Tier-0 CINECA system based on Lenovo NeXtScale platform. The current configuration of MARCONI consists of: • 3600 Intel Knights Landing nodes, each equipped with 1 Intel Xeon Phi 7250 @1.4 GHz, with 68 cores each and 96 GB of RAM, also named as MARCONI A2 – KNL • 3216 Intel SkyLake nodes, each equipped with 2 Intel Xeon 8160 @ 2.1 GHz, with 24 cores each and 192 GB of RAM, also named as MARCONI A3 - SKL 4 CANOPIE-HPC@SC19 www.mso4sc.eu | www.prace-ri.eu Containers 5 CANOPIE-HPC@SC19 www.mso4sc.eu | www.prace-ri.eu Native vs. Containers vs. VMs 6 CANOPIE-HPC@SC19 www.prace-ri.eu Why containers? ▶ Without containers: "I need software X, and here is the installation guide, please install it!” ▶ With containers: "I need software X, here is the name of its Docker image, please pull it" ▶ Very little performance degradation compared to native ▶ Security: SeLinux, Capability whitelist, syscall whitelist, and user namespaces 7 CANOPIE-HPC@SC19 www.prace-ri.eu What containers don’t solve ▶ Hardware architecture and kernel incompatibility. ▶ Operational maintenance mess (e.g. two different versions of MPI). ▶ Containers are not for huge software packages, e.g. Bio-Linux. To package those in one unit, VMs are more suitable 8 CANOPIE-HPC@SC19 www.prace-ri.eu MPI: Message Passing Interface 9 CANOPIE-HPC@SC19 www.mso4sc.eu | www.prace-ri.eu https://devblogs.nvidia.com/introduction-cuda-aware-mpi/ 10 CANOPIE-HPC@SC19 www.prace-ri.eu OSI Network model MPI https://medium.com/@int0x33/day-51-understanding-the-osi-model-f22d5f3df756 11 CANOPIE-HPC@SC19 www.prace-ri.eu General MPI program structure https://computing.llnl.gov/tutorials/mpi/ 12 CANOPIE-HPC@SC19 www.prace-ri.eu History ▶ May 1994: Final version of MPI-1.0 released ▶ MPI-1.1 (Jun 1995) ▶ MPI-1.2 (Jul 1997) ▶ MPI-1.3 (May 2008). ▶ 1998: MPI-2 picked up where the first MPI specification left off, and addressed topics which went far beyond the MPI-1 specification. ▶ MPI-2.1 (Sep 2008) ▶ MPI-2.2 (Sep 2009) ▶ Sep 2012: The MPI-3.0 standard was approved. ▶ MPI-3.1 (Jun 2015) ▶ Current: The MPI-4.0 standard is under development. https://computing.llnl.gov/tutorials/mpi/ 13 CANOPIE-HPC@SC19 www.prace-ri.eu Most pupular Distributions ▶ OpenMPI ▶ MPICH ▶ MVAPICH2 ▶ Intel MPI 14 CANOPIE-HPC@SC19 www.prace-ri.eu Common MPI Application Binary Interface (ABI) ▶ From Intel MPI 5.0 and MPICH 3.1 (and MVAPICH2 1.9 and higher which is based on MPICH 3.1), the libraries are interchangeable at the binary level, using common ABI. ▶ This in practice means that one can build the application with MPICH, but, run it using the Intel MPI libraries, and thus taking advantage of the Intel MPI functionality. https://software.intel.com/en-us/articles/using-intelr-mpi-library-50-with-mpich3-based-applications 15 CANOPIE-HPC@SC19 www.prace-ri.eu PMI: Process Management Interface PMI allows different process-management frameworks to interact with parallel programming libraries such as MPI Pavan Balaji, et. al. 2010. PMI: a scalable parallel process-management interface for extreme-scale systems. In Proceedings of the 17th European MPI users' group meeting conference on Recent advances in the message passing interface (EuroMPI'10), Rainer Keller, Edgar Gabriel, Michael 16 CANOPIE-HPC@SC19 www.prace-ri.eu Resch, and Jack Dongarra (Eds.). Springer-Verlag, Berlin, Heidelberg, 31-41. Container platform choice 17 CANOPIE-HPC@SC19 www.mso4sc.eu | www.prace-ri.eu Container platform choice Docker: ▶ Not designed for running HPC distributed applications. ▶ Relies on the Docker engine and daemon to be installed and running on each node. ▶ Each container has its own process which is fired up by the docker daemon. ▶ Containers run by default as root. While this can be avoided by using user namespaces, getting access to the main docker command, users can easily escape this limitation, which introduces a security risk. Each docker engine has its own image repository that, so far, cannot be shared with other docker engines. This means that running a parallel application X on e.g. 100 nodes, a X docker image need to be replicated 100 times, one per node. ▶ Docker so far has no support for MPI. 18 CANOPIE-HPC@SC19 www.prace-ri.eu Container platform choice Podman: ▶ Developed and maintained by Red Hat. ▶ A lightweight runtime for managing and running docker containers without the overhead of the docker engine and daemon. ▶ Not mainly designed for and is currently lacking support for HPC workloads. ▶ Podman can run rootless, but to to that (similar to charliecloud), it uses user namespaces ▶ Its rootless mode is not compatible with most parallel filesystems. 19 CANOPIE-HPC@SC19 www.prace-ri.eu Podman MPI example $ mpirun --hostfile hostfile \ --mca orte_tmpdir_base /tmp/podman-mpirun \ podman run --env-host \ -v /tmp/podman-mpirun:/tmp/podman-mpirun \ --userns=keep-id \ --net=host --pid=host --ipc=host \ quay.io/adrianreber/mpi-test:30 /home/ring Rank 2 has cleared MPI_Init Rank 2 has completed ring Rank 2 has completed MPI_Barrier Rank 3 has cleared MPI_Init Rank 3 has completed ring ….. https://podman.io/blogs/2019/09/26/podman-in-hpc.html 20 CANOPIE-HPC@SC19 www.prace-ri.eu Container platform choice Charliecloud: ▶ open source project which is based on a user-defined software stack (UDSS). ▶ Designed to manage and run docker based containers, and One important advantage that it is very lightweight. ▶ It enables running containers as the user, by implementing user namespaces. ▶ The main disadvantage is the requirement of a recent kernel (4.4 is recommended) that supports user namespaces. Many EU HPC clusters still have their compute nodes with old kernels (<3.10), i.e. don't support user namespaces which is a showstopper for charliecloud. ▶ Uses Slurm srun to run containers with MPI installed. 21 CANOPIE-HPC@SC19 www.prace-ri.eu Container platform choice Shifter: ▶ A platform to run Docker containers that is developed by the National Energy Research Scientific Computing Center (NERSC) and is deployed in production on a Slurm cluster of Cray supercomputers. ▶ Shifter uses its own image format to which both Docker images and VMs are converted. The Docker engine is replaced by image manager for managing the new formated images. ▶ Despite the fact that Shifter has proven to be useful managing HPC workflows, to use images in shifter, users must submit their images to a root controlled image gateway via a RESTful API. ▶ The main drawback of shifter is the lack of community ▶ Supports MPI depending on MPICH Application Binary Interface (ABI) 22 CANOPIE-HPC@SC19 www.prace-ri.eu Container platform choice Singularity: ▶ Singularity is the third and last HPC-specific container platform (after shifter and charliecloud). It is developed and maintained by Sylabs Inc. ▶ Unlike shifter, it doesn't have that complex architecture (with image manager and gateway) and can be simply installed as a environment module. ▶ Unlike Podman and charliecloud, it does not require kernel user namespaces to run rootless containers, and thus can be deployed on old Linux kernels. ▶ Singularity supports both Docker image format and its own native Single Image "flat" Format (SIF). ▶ Singularity is the most widely deployed container platform on HPC systems, currently there are 25,000+ systems running Singularity. ▶ Has build-in support for MPI (OpenMPI, MPICH, IntelMPI) 23 CANOPIE-HPC@SC19 www.prace-ri.eu Singularity Workflow 24 CANOPIE-HPC@SC19 www.mso4sc.eu | www.prace-ri.eu Singularity and MPI 25 CANOPIE-HPC@SC19 www.mso4sc.eu | www.prace-ri.eu Container platform choice Singularity: ▶ Data analysis example: Computing principal components for the first 10,000 variants from the 1000 Genomes Project chromosomes 21 and 22: wget https://…/chr21.head.vcf.gz wget https://…/chr22.head.vcf.gz LANG=C CHUNKSIZE=10000000 mpirun -x LANG -x CHUNKSIZE -np 2 singularity run -H $(pwd) variant_pca.img 26 CANOPIE-HPC@SC19 www.prace-ri.eu OpenMPI Singularity workflow ▶ From shell (or resource manager) mpirun gets called ▶ mpirun forks and exec ORTE (Open Run-Time Environment) daemon ▶ ORTED (ORTErun) process creates PMI ▶ ORTED forks == to the number of process per node requested ▶ ORTED children exec to original command passed to mpirun (Singularity) ▶ Each Singularity execs the command passed inside the given container ▶ Each MPI program links in the dynamic Open MPI libraries (ldd) ▶ OpenMPI libraries continue to open the non-ldd shared libraries (ldopen) ▶ OpenMPI libraries connect back to original ORTED via PMI ▶ All non-shared memory communication occurs through the PMI and then to local interfaces (e.g. InfiniBand) 27 CANOPIE-HPC@SC19 www.prace-ri.eu Problem Statement 28 CANOPIE-HPC@SC19 www.mso4sc.eu | www.prace-ri.eu
Recommended publications
  • Soluciones Para Entornos HPC
    Soluciones para entornos HPC Dr. Abel Francisco Paz Gallardo. IT Manager / Project Leader @ CETA-Ciemat [email protected] V Jornadas de Supercomputación y Avances en Tecnología CETA-Ciemat/ Noviembre 2012 Soluciones para entornos HPC Abel Francisco Paz Gallardo EXTREMADURA RESEARCH CENTER FOR ADVANCED TECHNOLOGIES INDICE 1 HPC… ¿Qué? ¿Cómo? . 2 Computación. (GPGPU,. UMA/NUMA,. etc.) 3 Almacenamiento en HPC . 4 Gestión de recursos . Soluciones para entornos HPC CETA-Ciemat/ Noviembre 2012 Soluciones para entornos HPC Abel Francisco Paz Gallardo 2 EXTREMADURA RESEARCH CENTER FOR ADVANCED TECHNOLOGIES 1 HPC… ¿Qué? ¿Cómo? ¿Qué? o HPC – High Performance Computing (Computación de alto rendimiento) o Objetivo principal: Resolución de problemas complejos ¿Cómo? Tres pilares fundamentales: o Procesamiento = Cómputo o Almacenamiento o Gestión de recursos HPC = + + CETA-Ciemat/ Noviembre 2012 Soluciones para entornos HPC Abel Francisco Paz Gallardo EXTREMADURA RESEARCH CENTER FOR ADVANCED TECHNOLOGIES 2 Computación (GPGPU, NUMA, etc.) ¿Qué es una GPU? Primera búsqueda en 2006: - Gas Particulate Unit Unidad de partículas de gas ¿? GPU = Graphics Processing Unit (Unidad de procesamiento gráfico). CETA-Ciemat/ Noviembre 2012 Soluciones para entornos HPC Abel Francisco Paz Gallardo EXTREMADURA RESEARCH CENTER FOR ADVANCED TECHNOLOGIES 2 Computación (GPGPU, NUMA, etc.) La cuestión es… Si una GPU en un videojuego procesa miles de polígonos, texturas y sombras en tiempo real… ¿Por qué no utilizar esta tecnología para procesamiento de datos? CETA-Ciemat/
    [Show full text]
  • The Component Architecture of Open Mpi: Enabling Third-Party Collective Algorithms∗
    THE COMPONENT ARCHITECTURE OF OPEN MPI: ENABLING THIRD-PARTY COLLECTIVE ALGORITHMS∗ Jeffrey M. Squyres and Andrew Lumsdaine Open Systems Laboratory, Indiana University Bloomington, Indiana, USA [email protected] [email protected] Abstract As large-scale clusters become more distributed and heterogeneous, significant research interest has emerged in optimizing MPI collective operations because of the performance gains that can be realized. However, researchers wishing to develop new algorithms for MPI collective operations are typically faced with sig- nificant design, implementation, and logistical challenges. To address a number of needs in the MPI research community, Open MPI has been developed, a new MPI-2 implementation centered around a lightweight component architecture that provides a set of component frameworks for realizing collective algorithms, point-to-point communication, and other aspects of MPI implementations. In this paper, we focus on the collective algorithm component framework. The “coll” framework provides tools for researchers to easily design, implement, and exper- iment with new collective algorithms in the context of a production-quality MPI. Performance results with basic collective operations demonstrate that the com- ponent architecture of Open MPI does not introduce any performance penalty. Keywords: MPI implementation, Parallel computing, Component architecture, Collective algorithms, High performance 1. Introduction Although the performance of the MPI collective operations [6, 17] can be a large factor in the overall run-time of a parallel application, their optimiza- tion has not necessarily been a focus in some MPI implementations until re- ∗This work was supported by a grant from the Lilly Endowment and by National Science Foundation grant 0116050.
    [Show full text]
  • HP-CAST 20 Final Agenda V3.0 (Sessions and Chairs Are Subject to Change Without Notice)
    HP Consortium for Advanced Scientific and Technical Computing World-Wide User Group Meeting ISS Hyperscale and HPC Organization The Westin Hotel Leipzig HP-CAST 20 Final Agenda V3.0 (Sessions and chairs are subject to change without notice) Supported by: Thursday, June 13th – Registration & Get-Together 18:00 – 22:00 Registration 18:00 – 22:00 Welcome Reception for All Attendees – Sky Bar FALCO at the Westin Hotel Friday, June 14th – Conference Section 08:00 – 18:00 Registration Board Introduction and HP Executive Updates HP-Liaison Frank Baetke, HP 08:15 – 09:20 HP-CAST President Rudolf Lohner, KIT/SCC HP Executive Updates on HPC Trends Paul Santeler, HP Scott Misage, HP Invited Partner Keynotes and Updates 09:20 – 09:50 High-Performance Computing Trends in the Georg Scheuerer, ANSYS Europe Manufacturing Industry 09:50 – 10:00 Making HPC Simulation Access Easier Benoît Vautrin, OVH (Oxalya) France 10:00 – 10:30 Break Invited Customer Keynotes 10:30 – 11:00 IT Infrastructure for HPC and Challenges of HPC Wolfgang Burke, BMW Germany Collocation at BMW 11:00 – 11:20 Deployment of HP Clusters in Iceland Guõbrandur R. Sigurõsson, Opin Kerfi Iceland 11:20 – 11:30 Q&A Session Short Processor Technology Updates Roberto Dognini, AMD Karl Freund, Calxeda (ARM) 11:30 – 12:30 John Hengeveld, Intel Timothy Lanfear, NVIDIA Sanjay Bhal, TI (ARM/DSP) 12:30 – 13:30 Lunch Friday, June 14th – Conference Section Attention: ATTENDANCE RESTRICTIONS APPLY Platforms and Products – #1 In Detail Roadmaps & Updates for Ed Turkel, HP 13:30 – 14:00 - Next Generation
    [Show full text]
  • Application-Level Fault Tolerance and Resilience in HPC Applications
    Doctoral Thesis Application-level Fault Tolerance and Resilience in HPC Applications Nuria Losada 2018 Application-level Fault Tolerance and Resilience in HPC Applications Nuria Losada Doctoral Thesis July 2018 PhD Advisors: Mar´ıaJ. Mart´ın Patricia Gonz´alez PhD Program in Information Technology Research Dra. Mar´ıaJos´eMart´ınSantamar´ıa Dra. Patricia Gonz´alezG´omez Profesora Titular de Universidad Profesora Titular de Universidad Dpto. de Ingenier´ıade Computadores Dpto. de Ingenier´ıade Computadores Universidade da Coru~na Universidade da Coru~na CERTIFICAN Que la memoria titulada \Application-level Fault Tolerance and Resilience in HPC Applications" ha sido realizada por D~na.Nuria Losada L´opez-Valc´arcelbajo nuestra direcci´onen el Departamento de Ingenier´ıade Computadores de la Universidade da Coru~na,y concluye la Tesis Doctoral que presenta para optar al grado de Doctora en Ingenier´ıaInform´aticacon la Menci´onde Doctor Internacional. En A Coru~na,a 26 de Julio de 2018 Fdo.: Mar´ıaJos´eMart´ınSantamar´ıa Fdo.: Patricia Gonz´alezG´omez Directora de la Tesis Doctoral Directora de la Tesis Doctoral Fdo.: Nuria Losada L´opez-Valc´arcel Autora de la Tesis Doctoral A todos los que lo hab´eishecho posible. Acknowledgments I would especially like to thank my advisors, Mar´ıaand Patricia, for all their support, hard work, and all the opportunities they've handed me. I consider my- self very lucky to have worked with them during these years. I would also like to thank Gabriel and Basilio for their collaboration and valuable contributions to the development of this work. I would like to say thanks to all my past and present colleagues in the Computer Architecture Group and in the Faculty of Informatics for their fellowship, support, and all the coffee breaks and dinners we held together.
    [Show full text]
  • Presentación De Powerpoint
    CENTRO DE SUPERCOMPUTACIÓN DE GALICIA High Scalability Multipole Method. Solving Half Billion of Unknowns J.C. Mouriño¹, A. Gómez¹, J.M. Taboada², L. Landesa², J.M. Bértolo³, F. Obelleiro³, J.L. Rodríguez³ ¹ Supercomputing Center of Galicia (CESGA), Spain ² University of Extremadura, Spain ³ University of Vigo, Spain International Supercomputing Conference Hamburg, June 23th 2009 Outline • Introduction • FMM-FFT algorithm • Parallel implementation • HEMCUVE. Challenge history • Experimental Results • Conclusions 2 Introduction • Numerical solution of the integro-differential electromagnetic equations -> important in industry • High frequency in large objects -> scalability limits – 0.78 m² at 3 GHz -> 15000 unknowns – real car at 79 GHz -> 400 million unknowns • Great effort reducing computational time of MoM • Fast Multipole Method (FMM) -> O(N³/²) good scalability • Multilevel FMM (MLFMA) -> O(N log N) poor scalability • Modern HPC systems have thousands of cores • FMM-FFT reduces complexity preserving scalability 3 Introduction • FMM – The geometry is partitioned into M groups using an oct-tree decomposition algorithm – Parallelization: • k-space samples are independent from one another • The workload can be partitioned in the k-space • MLFMA – The same algorithm can be recursively applied in a hierarchical multilevel oct-tree decomposition – Computational cost reduced to O(N log N) – Parallelization: • k-space samples are not independent due to interpolation/ anterpolation of fields across levels • Usually workload distributed by
    [Show full text]
  • Enabling the Deployment of Virtual Clusters on the VCOC Experiment of the Bonfire Federated Cloud
    CLOUD COMPUTING 2012 : The Third International Conference on Cloud Computing, GRIDs, and Virtualization Enabling the Deployment of Virtual Clusters on the VCOC Experiment of the BonFIRE Federated Cloud Raul Valin, Luis M. Carril, J. Carlos Mouri˜no, Carmen Cotelo, Andr´es G´omez, and Carlos Fern´andez Supercomputing Centre of Galicia (CESGA) Santiago de Compostela, Spain Email: rvalin,lmcarril,jmourino,carmen,agomez,[email protected] Abstract—The BonFIRE project has developed a federated OpenCirrus. BonFIRE offers an experimenter control of cloud that supports experimentation and testing of innovative available resources. It supports dynamically creating, updat- scenarios from the Internet of Services research community. ing, reading and deleting resources throughout the lifetime Virtual Clusters on federated Cloud sites (VCOC) is one of the supported experiments of the BonFIRE Project whose main of an experiment. Compute resources can be configured with objective is to evaluate the feasibility of using multiple Cloud application-specific contextualisation information that can environments to deploy services which need the allocation provide important configuration information to the virtual of a large pool of CPUs or virtual machines to a single machine (VM); this information is available to software user (as High Throughput Computing or High Performance applications after the machine is started. BonFIRE also Computing). In this work, we describe the experiment agent, a tool developed on the VCOC experiment to facilitate the supports elasticity within an experiment, i.e., dynamically automatic deployment and monitoring of virtual clusters on create, update and destroy resources from a running node of the BonFIRE federated cloud. This tool was employed in the experiment, including cross-testbed elasticity.
    [Show full text]
  • Efficient Multithreaded Untransposed, Transposed Or Symmetric Sparse
    Efficient Multithreaded Untransposed, Transposed or Symmetric Sparse Matrix-Vector Multiplication with the Recursive Sparse Blocks Format Michele Martonea,1,∗ aMax Planck Society Institute for Plasma Physics, Boltzmannstrasse 2, D-85748 Garching bei M¨unchen,Germany Abstract In earlier work we have introduced the \Recursive Sparse Blocks" (RSB) sparse matrix storage scheme oriented towards cache efficient matrix-vector multiplication (SpMV ) and triangular solution (SpSV ) on cache based shared memory parallel computers. Both the transposed (SpMV T ) and symmetric (SymSpMV ) matrix-vector multiply variants are supported. RSB stands for a meta-format: it recursively partitions a rectangular sparse matrix in quadrants; leaf submatrices are stored in an appropriate traditional format | either Compressed Sparse Rows (CSR) or Coordinate (COO). In this work, we compare the performance of our RSB implementation of SpMV, SpMV T, SymSpMV to that of the state-of-the-art Intel Math Kernel Library (MKL) CSR implementation on the recent Intel's Sandy Bridge processor. Our results with a few dozens of real world large matrices suggest the efficiency of the approach: in all of the cases, RSB's SymSpMV (and in most cases, SpMV T as well) took less than half of MKL CSR's time; SpMV 's advantage was smaller. Furthermore, RSB's SpMV T is more scalable than MKL's CSR, in that it performs almost as well as SpMV. Additionally, we include comparisons to the state-of-the art format Compressed Sparse Blocks (CSB) implementation. We observed RSB to be slightly superior to CSB in SpMV T, slightly inferior in SpMV, and better (in most cases by a factor of two or more) in SymSpMV.
    [Show full text]
  • Searching for Genetic Interactions in Complex Disease by Using Distance Correlation
    Searching for genetic interactions in complex disease by using distance correlation Fernando Castro-Prado, University and Health Research Institute of Santiago de Compostela, Spain. E-mail: [email protected] Javier Costas Health Research Institute of Santiago de Compostela, Spain. and Wenceslao González-Manteiga and David R. Penas University of Santiago de Compostela, Spain. Summary. Understanding epistasis (genetic interaction) may shed some light on the ge- nomic basis of common diseases, including disorders of maximum interest due to their high socioeconomic burden, like schizophrenia. Distance correlation is an association measure that characterises general statistical independence between random variables, not only the linear one. Here, we propose distance correlation as a novel tool for the detection of epistasis from case-control data of single nucleotide polymorphisms (SNPs). This approach will be developed both theoretically (mathematical statistics, in a context of high-dimensional statistical inference) and from an applied point of view (simulations and real datasets). Keywords: Association measures; Distance correlation; Epistasis; Genomics; High- dimensional statistical inference; Schizophrenia 1. Introduction The application field that motivates the present article is going to be explained here- inafter. The starting point is a genomic problem, whose importance and interest will be addressed. In addition, the state of the art on this field of knowledge will be sum- marised; underscoring one of the most recent techniques, which has a strong theoretical basis. Upon this, some hypotheses will be made. arXiv:2012.05285v1 [math.ST] 9 Dec 2020 1.1. Epistasis in complex disease The role of heredity in psychiatry has been studied for almost a century, with Pearson (1931) not having “the least hesitation” in asserting its relevance.
    [Show full text]
  • Infiniband and 10-Gigabit Ethernet for Dummies
    The MVAPICH2 Project: Latest Developments and Plans Towards Exascale Computing Presentation at OSU Booth (SC ‘19) by Hari Subramoni The Ohio State University E-mail: [email protected] http://www.cse.ohio-state.edu/~subramon Drivers of Modern HPC Cluster Architectures High Performance Accelerators / Coprocessors Interconnects - InfiniBand high compute density, high Multi-core <1usec latency, 200Gbps performance/watt SSD, NVMe-SSD, Processors Bandwidth> >1 TFlop DP on a chip NVRAM • Multi-core/many-core technologies • Remote Direct Memory Access (RDMA)-enabled networking (InfiniBand and RoCE) • Solid State Drives (SSDs), Non-Volatile Random-Access Memory (NVRAM), NVMe-SSD • Accelerators (NVIDIA GPGPUs and Intel Xeon Phi) • Available on HPC Clouds, e.g., Amazon EC2, NSF Chameleon, Microsoft Azure, etc. NetworkSummit Based Computing Sierra Sunway K - Laboratory OSU Booth SC’19TaihuLight Computer 2 Parallel Programming Models Overview P1 P2 P3 P1 P2 P3 P1 P2 P3 Memo Memo Memo Logical shared memory Memor Memor Memor Shared Memory ry ry ry y y y Shared Memory Model Distributed Memory Model Partitioned Global Address Space (PGAS) SHMEM, DSM MPI (Message Passing Interface)Global Arrays, UPC, Chapel, X10, CAF, … • Programming models provide abstract machine models • Models can be mapped on different types of systems – e.g. Distributed Shared Memory (DSM), MPI within a node, etc. • PGAS models and Hybrid MPI+PGAS models are Network Based Computing Laboratorygradually receiving importanceOSU Booth SC’19 3 Designing Communication Libraries for Multi-Petaflop and Exaflop Systems: Challenges Co-DesignCo-Design Application Kernels/Applications OpportuniOpportuni tiesties andand Middleware ChallengeChallenge ss acrossacross Programming Models VariousVarious MPI, PGAS (UPC, Global Arrays, OpenSHMEM), LayersLayers CUDA, OpenMP, OpenACC, Cilk, Hadoop (MapReduce), Spark (RDD, DAG), etc.
    [Show full text]
  • Enabling Rootless Linux Containers in Multi-User Environments: the Udocker Tool
    DESY 17-096 Enabling rootless Linux Containers in multi-user environments: the udocker tool Jorge Gomes1, Emanuele Bagnaschi2, Isabel Campos3, Mario David1, Lu´ısAlves1, Jo~aoMartins1, Jo~aoPina1, Alvaro L´opez-Garc´ıa3, and Pablo Orviz3 1Laborat´oriode Instrumenta¸c~aoe F´ısicaExperimental de Part´ıculas(LIP), Lisboa, Portugal 2Deutsches Elektronen-Synchrotron (DESY), 22607 Hamburg, Germany 3IFCA, Consejo Superior de Investigaciones Cient´ıficas-CSIC,Santander, Spain June 5, 2018 Abstract Containers are increasingly used as means to distribute and run Linux services and applications. In this paper we describe the architectural design and implementation of udocker, a tool which enables the user to execute Linux containers in user mode. We also present a few practical applications, using a range of scientific codes characterized by different requirements: from single core execution to MPI parallel execution and execution on GPGPUs. 1 Introduction Technologies based on Linux containers have become very popular among soft- ware developers and system administrators. The main reason behind this suc- cess is the flexibility and efficiency that containers offer when it comes to pack- ing, deploying and running software. A given software can be containerized together with all its dependencies in arXiv:1711.01758v2 [cs.SE] 1 Jun 2018 such a way that it can be seamlessly executed regardless of the Linux distribution used by the designated host systems. This is achieved by using advanced features of modern Linux kernels [1], namely control groups and namespaces isolation [2, 3]. Using both features, a set of processes can be placed in a fully isolated environment (using namespaces isolation), with a given amount of resources, such as CPU or RAM, allocated to it (using control groups).
    [Show full text]
  • User Manual Revision: C96e096
    Bright Cluster Manager 8.1 User Manual Revision: c96e096 Date: Fri Sep 14 2018 ©2018 Bright Computing, Inc. All Rights Reserved. This manual or parts thereof may not be reproduced in any form unless permitted by contract or by written permission of Bright Computing, Inc. Trademarks Linux is a registered trademark of Linus Torvalds. PathScale is a registered trademark of Cray, Inc. Red Hat and all Red Hat-based trademarks are trademarks or registered trademarks of Red Hat, Inc. SUSE is a registered trademark of Novell, Inc. PGI is a registered trademark of NVIDIA Corporation. FLEXlm is a registered trademark of Flexera Software, Inc. PBS Professional, PBS Pro, and Green Provisioning are trademarks of Altair Engineering, Inc. All other trademarks are the property of their respective owners. Rights and Restrictions All statements, specifications, recommendations, and technical information contained herein are current or planned as of the date of publication of this document. They are reliable as of the time of this writing and are presented without warranty of any kind, expressed or implied. Bright Computing, Inc. shall not be liable for technical or editorial errors or omissions which may occur in this document. Bright Computing, Inc. shall not be liable for any damages resulting from the use of this document. Limitation of Liability and Damages Pertaining to Bright Computing, Inc. The Bright Cluster Manager product principally consists of free software that is licensed by the Linux authors free of charge. Bright Computing, Inc. shall have no liability nor will Bright Computing, Inc. provide any warranty for the Bright Cluster Manager to the extent that is permitted by law.
    [Show full text]
  • Prace6ip-D7.1
    E-Infrastructures H2020- INFRAEDI-2018-2020 INFRAEDI-01-2018: Pan-European High Performance Computing infrastructure and services (PRACE) PRACE-6IP PRACE Sixth Implementation Phase Project Grant Agreement Number: INFRAEDI-823767 D7.1 Periodic Report on Results from Projects Supported by Applications Enabling and Porting Services Final Version: 1.0 Author(s): Gabriel Hautreux, CINES; Chris Johnson, EPCC; Sebastian Lührs, JUELICH Date: 16.10.2020 D7.1 Periodic Report on Results from Projects Supported by Applications Enabling and Porting Services Project and Deliverable Information Sheet PRACE Project Project Ref. №: INFRAEDI-823767 Project Title: PRACE Sixth Implementation Phase Project Project Web Site: http://www.prace-project.eu Deliverable ID: D7.1 Deliverable Nature: Report Dissemination Level: Contractual Date of Delivery: PU* 30 / October / 2020 Actual Date of Delivery: 30 / October / 2020 EC Project Officer: Leonardo Flores Añover * - The dissemination level are indicated as follows: PU – Public, CO – Confidential, only for members of the consortium (including the Commission Services) CL – Classified, as referred to in Commission Decision 2005/444/EC. Document Control Sheet Title: Periodic Report on Results from Projects Supported by Document Applications Enabling and Porting Services ID: D7.1 Version: <1.0> Status: Final Available at: http://www.prace-project.eu Software Tool: Microsoft Word 2016 File(s): D7.1.docx Written by: Gabriel Hautreux, CINES; Chris Johnson, Authorship EPCC; Sebastian Lührs, JUELICH Contributors: Adam
    [Show full text]