Commodity Cluster Using Single System Image Based on Linux/Kerrighed for High-Performance Computing

Total Page:16

File Type:pdf, Size:1020Kb

Commodity Cluster Using Single System Image Based on Linux/Kerrighed for High-Performance Computing Proc. of 2016 3rd Int. Conf. on Information Tech., Computer, and Electrical Engineering (ICITACEE), Oct 19-21st, 2016, Semarang, Indonesia Commodity Cluster Using Single System Image Based on Linux/Kerrighed for High-Performance Computing Iwan Setiawan and Eko Murdyantoro Department of Electrical Engineering Universitas Jenderal Soedirman Purwokerto, Indonesia {stwn, eko.murdyantoro}@unsoed.ac.id Abstract—Commodity Information Technology (IT) applications with 16 PC-based homogeneous nodes. It has been infrastructure, including hardware, software, and networking, is inspiring academics, researchers, and practitioners to build commonly used in computing facilities, such as computer their own clusters for HPC with commodity infrastructure, laboratory. Unfortunately, the infrastructure in that particular including software that is commonly Free/Libre/Open Source place need to be updated and upgraded regularly to cater to the Software (FLOSS). For instance, Linux for the OS and MPICH demands of users and applications for more and more computing for the parallel library, one of the implementations of Message resources, especially in scientific computation. The existing Passing Interface (MPI). infrastructure can be utilized to tackle the issue by forming it into cluster computing. We report our exploration results in Many organizations, specifically universities and research constructing Vincster, a low-cost and scalable commodity cluster institutes, employ commodity IT infrastructure in their for high-performance computing (HPC) to utilize the existing IT computing facilities, e.g. computer laboratory. One of the infrastructure in a computer laboratory. The cluster uses one issues in computer laboratory is how to keep up with the additional computer to serve as a head node containing a Single increasingly requests of more computing resources for users System Image (SSI) Operating System (OS) based on Linux with and their applications, especially in scientific computation. Kerrighed as the kernel-level SSI support, along with Hence, the infrastructure in that site need to be updated and middleware and application softwares that build up the SSI upgraded regularly to meet the demands. The existing approach. Vincster has been constructed and implemented with 7 infrastructure can be utilized to overcome the issue by forming SSI key services/features utilizing 17 compute nodes from 20 it into scalable commodity cluster for HPC at minimal cost. available computers. We evaluated computational performance of the cluster by running parallel matrix multiplication program. However, there is a challenge in building cluster computing with traditional method, i.e., to administer the cluster manually Keywords—commodity cluster; high-performance computing; or with tools including installation and maintenance, since the single system image; linux; kerrighed; cots; computer laboratory compute nodes with their computing resources are distributed. It is not easy to make the nodes coherent and identical to one I. INTRODUCTION another, and disk replication technique that applies across Nowadays, the need of computing resources is increasingly nodes is not flexible. Thus, users and applications still see demanding because of the big amount of data to be handled cluster as a system with distributed resources, not as a single and more CPU cycles to be used. Barroso, Dean, and Hӧ lzle and unified system. [1] stated that “on average, a single query on Google reads Single System Image (SSI), “the property of a system that hundreds of megabytes of data and consumes tens of billions of hides the distributed and heterogeneous of the available CPU cycles.” This trend coupled with the high cost and low resources and presents them to users and applications as a accessibility of traditional supercomputers, along with current single unified computing resource” as described by Buyya, advancement of high-performance microprocessors, high-speed Cortes, and Jin [5], is an approach to tackle the issue networks, and standard tools for HPC, makes commodity aforestated. By abstracting the distributed resources, the cluster cluster, cluster computing that uses fully commodity will be easier to use, administer, and maintain. SSI can be components or commodity-off-the-self (COTS) [2] such as implemented at different levels of abstraction: hardware, Personal Computer (PC), appealing solution for cost-effective OS/kernel, middleware, and application [3] [5] [6]. Each parallel computing [3]. The increasing need of computing implementation level has its advantages and disadvantages. resources trend itself is not just for commercial applications but The more low level the implementation, the more difficult the also computational science applications [3]. implementation has to be constructed, but with more Beowulf is an example of commodity cluster [2], which transparency to the higher levels of the system layer. was first built by Sterling et al. [4] for Earth and space science Cooperation between levels is needed to build a good SSI [5]. 978-1-5090-0890-2/16/$31.00 c 2016 IEEE 367 There is a study of three implementations of SSI OS based the cluster should include important or key SSI on Linux kernel, namely openMosix, OpenSSI, and Kerrighed. services/features in [5]. We attempt to incorporate any This study was conducted by Lottiaux, Gallard, Vallée, Morin, accessible methods to include the features into the cluster. The and Boissinot [7]. According to the study, Kerrighed offers the second one is the cluster need to be scalable in terms of best performance, notably regarding its Inter-Process flexibility for nodes to join or disjoin to the cluster (loosely Communication (IPC) and filesystem. Kerrighed mainly coupled system). The third one is the cluster should be low-cost targeted in kernel-level SSI providing the most transparent SSI, by minimizing efforts in buying, changing, and reconfiguring compared to other patchset-based implementations [6]. It also existing commodity IT infrastructure. has interesting features, e.g. cluster-wide process management, support for cluster-wide shared memory, cluster filesystem, Furthermore, there are environment characteristics in transparent process checkpointing, high-availability for user computer laboratory that should be addressed, related to the applications, and customizable SSI features. It supports several construction of the cluster. These characteristics include hardware architectures, i386/x86-32/IA32 for version <= 2.3.0, number of compute nodes and the heterogeneity of their and x86-64 for version >=2.4.0 [8]. specifications, including the types and number of cores/processors on each node. As a case study, we conducted Armay, Zulfikar, and Simaremare [9] have built a this research in a computer laboratory of Department of commodity cluster with 9 PC-based homogeneous nodes in a Engineering (now Faculty of Engineering), Universitas computer laboratory with traditional method. There is no single Jenderal Soedirman (Unsoed). We started this research by and unified-view of the cluster, since they do not use SSI doing survey of the existing commodity infrastructure in that approach. Moreover, they demonstrate a test using MPICH1 to laboratory to formulate the design and develop system proof that parallel processing can be run on the cluster, but architecture of the cluster. there is no computational performance presented in the study. Sandhya and Raju [10] built a commodity cluster using SSI Lastly, we use “Vincster” as the name of the cluster that approach based on Linux/Kerrighed with four PC-based nodes. stands for “vintage commodity cluster.” It represents our hope These nodes have two different specifications in processor and of this cluster’s goal, that is bringing back to life outdated memory. They report evaluation results of computational computing resources to harvest more its values. performance for 10 to 100 procesess computing the value of pi on the cluster, but they do not explain how those processes run III. SYSTEM DESCRIPTION on each of the compute nodes. In the study, the relation In this section, we describe the system architecture of between performance of the cluster, number of compute nodes Vincster, and also the cost needed for its construction. It and their cores/processors, along with various problem sizes, is consists of a head node, and a number of compute nodes. The not described. head node provides network services for the compute nodes, We attempt to explore on constructing cluster that addresses including network booting via Preboot Execution Environment the issues aforementioned. To summarize, our contributions are (PXE), Dynamic Host Configuration Protocol (DHCP) for as follows. assigning IP addresses dynamically, Trivial File Transfer Protocol (TFTP) for transfering kernel/OS image to compute 1. We revisit Beowulf/commodity cluster for HPC combined nodes, Network File System (NFS) for provisioning the root with SSI approach to address the issues by designing and filesystem and sharing storage that are needed to run full- constructing Vincster, a low-cost and scalable cluster based fledged SSI OS on compute nodes. on Linux OS with Kerrighed as the kernel-level SSI support together with middleware and application softwares that We use network booting and centralized storage technique build up the SSI approach. As a case study, we implement in Vincster. All PCs in the laboratory are booted by loading the design and deploy the commodity cluster over existing PXE first from ROM/BIOS or flash drive.
Recommended publications
  • Administració De Sistemes GNU Linux Mòdul4 Administració
    Administració local Josep Jorba Esteve PID_00238577 GNUFDL • PID_00238577 Administració local Es garanteix el permís per a copiar, distribuir i modificar aquest document segons els termes de la GNU Free Documentation License, Version 1.3 o qualsevol altra de posterior publicada per la Free Software Foundation, sense seccions invariants ni textos de la oberta anterior o posterior. Podeu consultar els termes de la llicència a http://www.gnu.org/licenses/fdl-1.3.html. GNUFDL • PID_00238577 Administració local Índex Introducció.................................................................................................. 5 1. Eines bàsiques per a l'administrador........................................... 7 1.1. Eines gràfiques i línies de comandes .......................................... 8 1.2. Documents d'estàndards ............................................................. 10 1.3. Documentació del sistema en línia ............................................ 13 1.4. Eines de gestió de paquets .......................................................... 15 1.4.1. Paquets TGZ ................................................................... 16 1.4.2. Fedora/Red Hat: paquets RPM ....................................... 19 1.4.3. Debian: paquets DEB ..................................................... 24 1.4.4. Nous formats d'empaquetat: Snap i Flatpak .................. 28 1.5. Eines genèriques d'administració ................................................ 36 1.6. Altres eines .................................................................................
    [Show full text]
  • Storage Administration Guide Storage Administration Guide SUSE Linux Enterprise Server 12 SP4
    SUSE Linux Enterprise Server 12 SP4 Storage Administration Guide Storage Administration Guide SUSE Linux Enterprise Server 12 SP4 Provides information about how to manage storage devices on a SUSE Linux Enterprise Server. Publication Date: September 24, 2021 SUSE LLC 1800 South Novell Place Provo, UT 84606 USA https://documentation.suse.com Copyright © 2006– 2021 SUSE LLC and contributors. All rights reserved. Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or (at your option) version 1.3; with the Invariant Section being this copyright notice and license. A copy of the license version 1.2 is included in the section entitled “GNU Free Documentation License”. For SUSE trademarks, see https://www.suse.com/company/legal/ . All other third-party trademarks are the property of their respective owners. Trademark symbols (®, ™ etc.) denote trademarks of SUSE and its aliates. Asterisks (*) denote third-party trademarks. All information found in this book has been compiled with utmost attention to detail. However, this does not guarantee complete accuracy. Neither SUSE LLC, its aliates, the authors nor the translators shall be held liable for possible errors or the consequences thereof. Contents About This Guide xii 1 Available Documentation xii 2 Giving Feedback xiv 3 Documentation Conventions xiv 4 Product Life Cycle and Support xvi Support Statement for SUSE Linux Enterprise Server xvii • Technology Previews xviii I FILE SYSTEMS AND MOUNTING 1 1 Overview
    [Show full text]
  • Ada Departmental Supercomputer Shared Memory GPU Cluster
    Ada Departmental Supercomputer Shared Memory GPU Cluster The Ada Departmental Supercomputer is designed to provide System Specifications near top 500 class supercomputing capabilities at your office Processors: Head Node: 2 AMD EPYC 7702 Processors or lab. (64 core-2.0/3.3 GHz) Compute Nodes: 1 AMD EPYC 7702P Proces- Ada is a hybrid supercomputer consisting of a large memory sor (64 core-2.2/3.2 GHz), 8 AMD Radeon head node and 2 to 5 compute nodes, each with eight AMD Instinct MI50 GPUs Radeon Instinct MI50 GPUs. With 5 compute nodes Ada con- Global Memory: 2TB or 4TB 3200 MHz DDR4 tains 448 AMD EPYC processor cores, 40 MI50 GPUs and 2 or 4 TB of globally shared memory. The compute nodes are Compute Node 128 GB 3200 MHz DDR4 (each) Memory: connected to the head node with 200 Gb/s Mellanox Infini- band. The Ada departmental supercomputer can be config- Storage: 1TB on-board M.2 OS SSD ured to deliver 1060 TFLOPS of FP16, 532 TFLOPS of FP32 12x 3.5" SATA/SAS hot-swap and 264 TFLOPS of FP64 GPU floating point performance SSD/HDD bays (head node) Additional 8x 2.5” SSD hot-swap bays on each capable of operating on large computational models. compute node Ada is a true symmetric multi-processing (SMP) computer Interconnect: ConnectX-6 VPI 200 Gb/s InfiniBand Dual Port with a large shared memory and a single operating system PCIe Gen 4 Host Bus Adapters user interface based on Centos 8 Linux. It provides a 1TB (No InfiniBand switch is needed) globally shared fast file system, and a large disk storage ar- I/O: 2x 1 Gb/s LAN ports ray.
    [Show full text]
  • Sprite File System There Are Three Important Aspects of the Sprite ®Le System: the Scale of the System, Location-Transparency, and Distributed State
    Naming, State Management, and User-Level Extensions in the Sprite Distributed File System Copyright 1990 Brent Ballinger Welch CHAPTER 1 Introduction ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ This dissertation concerns network computing environments. Advances in network and microprocessor technology have caused a shift from stand-alone timesharing systems to networks of powerful personal computers. Operating systems designed for stand-alone timesharing hosts do not adapt easily to a distributed environment. Resources like disk storage, printers, and tape drives are not concentrated at a single point. Instead, they are scattered around the network under the control of different hosts. New operating system mechanisms are needed to handle this sort of distribution so that users and application programs need not worry about the distributed nature of the underlying system. This dissertation explores the approach of centering a distributed computing environment around a shared network ®le system. The ®le system is chosen as a starting point because it is a heavily used service in stand-alone systems, and the read/write para- digm of the ®le system is a familiar one that can be applied to many system resources. The ®le system described in this dissertation provides a distributed name space for sys- tem resources, and it provides remote access facilities so all resources are available throughout the network. Resources accessible via the ®le system include disk storage, other types of peripheral devices, and user-implemented service applications. The result- ing system is one where resources are named and accessed via the shared ®le system, and the underlying distribution of the system among a collection of hosts is not important to users.
    [Show full text]
  • USB Composite Gadget Using CONFIG-FS on Dra7xx Devices
    Application Report SPRACB5–September 2017 USB Composite Gadget Using CONFIG-FS on DRA7xx Devices RaviB ABSTRACT This application note explains how to create a USB composite gadget, network control model (NCM) and abstract control model (ACM) from the user space using Linux® CONFIG-FS on the DRA7xx platform. Contents 1 Introduction ................................................................................................................... 2 2 USB Composite Gadget Using CONFIG-FS ............................................................................. 3 3 Creating Composite Gadget From User Space.......................................................................... 4 4 References ................................................................................................................... 8 List of Figures 1 Block Diagram of USB Composite Gadget............................................................................... 3 2 Selection of CONFIGFS Through menuconfig........................................................................... 4 3 Select USB Configuration Through menuconfig......................................................................... 4 4 Composite Gadget Configuration Items as Files and Directories ..................................................... 5 5 VID, PID, and Manufacturer String Configuration ....................................................................... 6 6 Kernel Logs Show Enumeration of USB Composite Gadget by Host ................................................ 6 7 Ping
    [Show full text]
  • Oracle® Linux 7 Managing File Systems
    Oracle® Linux 7 Managing File Systems F32760-07 August 2021 Oracle Legal Notices Copyright © 2020, 2021, Oracle and/or its affiliates. This software and related documentation are provided under a license agreement containing restrictions on use and disclosure and are protected by intellectual property laws. Except as expressly permitted in your license agreement or allowed by law, you may not use, copy, reproduce, translate, broadcast, modify, license, transmit, distribute, exhibit, perform, publish, or display any part, in any form, or by any means. Reverse engineering, disassembly, or decompilation of this software, unless required by law for interoperability, is prohibited. The information contained herein is subject to change without notice and is not warranted to be error-free. If you find any errors, please report them to us in writing. If this is software or related documentation that is delivered to the U.S. Government or anyone licensing it on behalf of the U.S. Government, then the following notice is applicable: U.S. GOVERNMENT END USERS: Oracle programs (including any operating system, integrated software, any programs embedded, installed or activated on delivered hardware, and modifications of such programs) and Oracle computer documentation or other Oracle data delivered to or accessed by U.S. Government end users are "commercial computer software" or "commercial computer software documentation" pursuant to the applicable Federal Acquisition Regulation and agency-specific supplemental regulations. As such, the use, reproduction, duplication, release, display, disclosure, modification, preparation of derivative works, and/or adaptation of i) Oracle programs (including any operating system, integrated software, any programs embedded, installed or activated on delivered hardware, and modifications of such programs), ii) Oracle computer documentation and/or iii) other Oracle data, is subject to the rights and limitations specified in the license contained in the applicable contract.
    [Show full text]
  • Unionfs: User- and Community-Oriented Development of a Unification File System
    Unionfs: User- and Community-Oriented Development of a Unification File System David Quigley, Josef Sipek, Charles P. Wright, and Erez Zadok Stony Brook University {dquigley,jsipek,cwright,ezk}@cs.sunysb.edu Abstract If a file exists in multiple branches, the user sees only the copy in the higher-priority branch. Unionfs allows some branches to be read-only, Unionfs is a stackable file system that virtually but as long as the highest-priority branch is merges a set of directories (called branches) read-write, Unionfs uses copy-on-write seman- into a single logical view. Each branch is as- tics to provide an illusion that all branches are signed a priority and may be either read-only writable. This feature allows Live-CD develop- or read-write. When the highest priority branch ers to give their users a writable system based is writable, Unionfs provides copy-on-write se- on read-only media. mantics for read-only branches. These copy- on-write semantics have lead to widespread There are many uses for namespace unifica- use of Unionfs by LiveCD projects including tion. The two most common uses are Live- Knoppix and SLAX. In this paper we describe CDs and diskless/NFS-root clients. On Live- our experiences distributing and maintaining CDs, by definition, the data is stored on a read- an out-of-kernel module since November 2004. only medium. However, it is very convenient As of March 2006 Unionfs has been down- for users to be able to modify the data. Uni- loaded by over 6,700 unique users and is used fying the read-only CD with a writable RAM by over two dozen other projects.
    [Show full text]
  • ODROID-HC2: 3.5” High Powered Storage  February 1, 2018
    ODROID WiFi Access Point: Share Files Via Samba February 1, 2018 How to setup an ODROID with a WiFi access point so that an ODROID’s hard drive can be accessed and modied from another computer. This is primarily aimed at allowing access to images, videos, and log les on the ODROID. ODROID-HC2: 3.5” High powered storage February 1, 2018 The ODROID-HC2 is an aordable mini PC and perfect solution for a network attached storage (NAS) server. This device home cloud-server capabilities centralizes data and enables users to share and stream multimedia les to phones, tablets, and other devices across a network. It is an ideal tool for many use Using SquashFS As A Read-Only Root File System February 1, 2018 This guide describes the usage of SquashFS PiFace: Control and Display 2 February 1, 2018 For those who have the PiFace Control and Display 2, and want to make it compatible with the ODROID-C2 Android Gaming: Data Wing, Space Frontier, and Retro Shooting – Pixel Space Shooter February 1, 2018 Variations on a theme! Race, blast into space, and blast things into pieces that are racing towards us. The fun doesn’t need to stop when you take a break from your projects. Our monthly pick on Android games. Linux Gaming: Saturn Games – Part 1 February 1, 2018 I think it’s time we go into a bit more detail about Sega Saturn for the ODROID-XU3/XU4 Gaming Console: Running Your Favorite Games On An ODROID-C2 Using Android February 1, 2018 I built a gaming console using an ODROID-C2 running Android 6 Controller Area Network (CAN) Bus: Implementation
    [Show full text]
  • Deploying OFS Technology in the Wild: a Case Study
    13th ANNUAL WORKSHOP 2017 DEPLOYING OFS TECHNOLOGY IN THE WILD A CASE STUDY Susan Coulter / HPC-Design Los Alamos National Laboratory [ March 31, 2017 ] LA-UR-17-22449 HOW THE STORY STARTS… . LANL / CSCNSI • Summer school for Junior/Senior Computer Science majors • Project: Compare 100G Ethernet to IB EDR • Cluster built with IB FDR • Preliminary test compared FDR to EDR 2 OpenFabrics Alliance Workshop 2017 FIRST WRINKLE . LANL deployed Damselfly IB backbone HUNTER • Only EDR systems in production • SM, slipknot cluster, redcap cluster • Most other systems FDR-connected • Built early with Mellanox-OFED GARCIA • Replaced with TOSS(RedHat) bundled OFS Trinity Lustre • Tri-Lab Operating System Stack Common Lustre • TOSS2 -> RedHat6 • TOSS3 -> RedHat7 • LANL upgrade schedule slower than LLNL upgrade schedule • LANL running version(s) LLNL has frozen Mid/Long Term Archive IB EDR 3 OpenFabrics Alliance Workshop 2017 WRINKLES WITHIN WRINKLES . Disk-ful / Disk-less / Configuration Management • Install / test Mellanox OFED on TOSS standalone system – easy • Non-standard kernels use Mellanox script – easy • Cfengine controls cluster configuration • RPMs only – automation preferred except under extreme circumstances • Local updates repo (kernel RPMs and associated libraries) • Newer version number • depmod –a » /etc/depmod.d/mlnx-ofa_kernel.conf • Hybrid images – RAM and NFS mount • Necessary kernel modules need to be in RAM » rdma_cm requires configfs.ko override ib_uverbs * weak-updates/mlnx-ofa_kernel/drivers/infiniband/core override ib_addr * weak-updates/mlnx-ofa_kernel/drivers/infiniband/core override ib_umad * weak-updates/mlnx-ofa_kernel/drivers/infiniband/core override ib_core * weak-updates/mlnx-ofa_kernel/drivers/infiniband/core 4 OpenFabrics Alliance Workshop 2017 SUCCESS ! . Campaign / Scality system upgraded • ~25% increase in performance • Uses >lots< of small messages .
    [Show full text]
  • Containers: a Sound Basis for a True Single System Image
    Containers : A Sound Basis For a True Single System Image Renaud Lottiaux, Christine Morin To cite this version: Renaud Lottiaux, Christine Morin. Containers : A Sound Basis For a True Single System Image. [Research Report] RR-4085, INRIA. 2000. inria-00072548 HAL Id: inria-00072548 https://hal.inria.fr/inria-00072548 Submitted on 24 May 2006 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. INSTITUT NATIONAL DE RECHERCHE EN INFORMATIQUE ET EN AUTOMATIQUE Containers : A Sound Basis For a True Single System Image Renaud Lottiaux, Christine Morin N˚4085 Novembre 2000 THÈME 1 apport de recherche ISRN INRIA/RR--4085--FR+ENG ISSN 0249-6399 Containers : A Sound Basis For a True Single System Image Renaud Lottiaux , Christine Morin Thème 1 — Réseaux et systèmes Projet PARIS Rapport de recherche n˚4085 — Novembre 2000 — 19 pages Abstract: Clusters of SMPs are attractive for executing shared memory parallel appli- cations but reconciling high performance and ease of programming remains an open issue. A possible approach is to provide an efficient Single System Image (SSI) operating system giving the illusion of an SMP machine. In this paper, we introduce the concept of container as a mechanism to unify global resource management at the lowest operating system level.
    [Show full text]
  • A Single System Image Java Operating System for Sensor Networks
    A SINGLE SYSTEM IMAGE JAVA OPERATING SYSTEM FOR SENSOR NETWORKS Emin Gun Sirer Rimon Barr John C. Bicket Daniel S. Dantas Computer Science Department Cornell University Ithaca, NY 14853 {egs, barr, bicket, ddantas}@cs.cornell.edu Abstract In this paper we describe the design and implementation of a distributed operating system for sensor net- works. The goal of our system is to extend total system lifetime through power-aware adaptation for sensor networking applications. Our system achieves this goal by providing a single system image of a unified Java virtual machine to applications over an ad hoc collection of heterogeneous sensors. It automatically and transparently partitions applications into components and dynamically finds a placement of these components on nodes within the sensor network to reduce energy consumption and increase system longevity. This paper describes the design and implementation of our system and examines the question of where and when to mi- grate components in a sensor network. We evaluate two practical, power-aware, general-purpose algorithms for object placement, as well as an adaptive scheme for deciding the time granularity of object migration. We demonstrate that our algorithms can increase sensor network longevity by a factor of four to five by effec- tively distributing energy consumption and avoiding hotspots. 1. Introduction able to components at each node, in particular the available power and bandwidth may change over Sensor networks simultaneously promise a radi- time and necessitate the relocation of application cally new class of applications and pose signifi- components. Further, event sources that are being cant challenges for application development.
    [Show full text]
  • View Article(3467)
    Problems of information technology, 2018, №1, 92–97 Kamran E. Jafarzade DOI: 10.25045/jpit.v09.i1.10 Institute of Information Technology of ANAS, Baku, Azerbaijan [email protected] COMPARATIVE ANALYSIS OF THE SOFTWARE USED IN SUPERCOMPUTER TECHNOLOGIES The article considers the classification of the types of supercomputer architectures, such as MPP, SMP and cluster, including software and application programming interfaces: MPI and PVM. It also offers a comparative analysis of software in the study of the dynamics of the distribution of operating systems (OS) for the last year of use in supercomputer technologies. In addition, the effectiveness of the use of CentOS software on the scientific network "AzScienceNet" is analyzed. Keywords: supercomputer, operating system, software, cluster, SMP-architecture, MPP-architecture, MPI, PVM, CentOS. Introduction Supercomputer is a computer with high computing performance compared to a regular computer. Supercomputers are often used for scientific and engineering applications that need to process very large databases or perform a large number of calculations. The performance of a supercomputer is measured in floating-point operations per second (FLOPS) instead of millions of instructions per second (MIPS). Since 2015, the supercomputers performing up to quadrillion FLOPS have started to be developed. Modern supercomputers represent a large number of high performance server computers, which are interconnected via a local high-speed backbone to achieve the highest performance [1]. Supercomputers were originally introduced in the 1960s and bearing the name or monogram of the companies such as Seymour Cray of Control Data Corporation (CDC), Cray Research over the next decades. By the end of the 20th century, massively parallel supercomputers with tens of thousands of available processors started to be manufactured.
    [Show full text]