Uva-DARE (Digital Academic Repository)

Total Page:16

File Type:pdf, Size:1020Kb

Uva-DARE (Digital Academic Repository) UvA-DARE (Digital Academic Repository) Scalable distributed data structures for database management Karlsson, S.J. Publication date 2000 Document Version Final published version Link to publication Citation for published version (APA): Karlsson, S. J. (2000). Scalable distributed data structures for database management. General rights It is not permitted to download or to forward/distribute the text or part of it without the consent of the author(s) and/or copyright holder(s), other than for strictly personal, individual use, unless the work is under an open content license (like Creative Commons). Disclaimer/Complaints regulations If you believe that digital publication of certain material infringes any of your rights or (privacy) interests, please let the Library know, stating your reasons. In case of a legitimate complaint, the Library will make the material inaccessible and/or remove it from the website. Please Ask the Library: https://uba.uva.nl/en/contact, or a letter to: Library of the University of Amsterdam, Secretariat, Singel 425, 1012 WP Amsterdam, The Netherlands. You will be contacted as soon as possible. UvA-DARE is a service provided by the library of the University of Amsterdam (https://dare.uva.nl) Download date:11 Oct 2021 atabasee Management Scalablee Distributed Data Structures s for r Databasee Management Academischh Proefschrift terr verkrijging van de graad van doctor aann de Universiteit van Amsterdam opp gezag van de Rector Magnificus prof.. dr. J. J. M. Franse tenn overstaan van een door het collegee voor promoties ingestelde commissie, inn het openbaar te verdedigen inn de Aula der Universiteit opp donderdag 14 december 2000, te 11.00 uur doorr Jonas S Karlsson geborenn te Enköping, Sweden Promotor:: Prof. Dr. M. L. Kersten Faculteit:: Faculteit der Natuurwetenschappen, Wiskunde en Inform; Thee research reported in this thesis has been partially carried out at thee University of Linköping, Sweden, within the Engineering Database Lab,, a research group at the Department of Computer and Information Sciencee at Linköping Institute of Technology. Thee research reported in this thesis has been partially carried out at CWI,, the Dutch national research laboratory for mathematics and com- puterr science, within the theme Data Mining and Knowledge Discovery, aa subdivision of the research cluster Information Systems. Thee research reported in this thesis has been carried out under the auspi- cess of SIKS, the Dutch Graduate School for Information and Knowledge Systems.. SIKS Dissertation Series No 2000-11 ISBNN 90 6196 498 9 Contents s II Scalable Distributed Data Structures 19 11 Preliminaries 21 1.11 Birthground of SDDSs 21 1.22 SDDSs 22 1.33 Requirements from SDDSs 22 1.44 Data Structures - Background 23 1.4.11 Retrieval Methods 23 1.4.22 Reasonable Properties 24 1.4.33 Basic Data Organization 25 1.4.44 Memory Management/Heaps 26 1.4.55 Linked Lists 27 1.4.66 Chained/Closed Hashing 27 1.4.77 Trees 27 1.4.88 Signature Files 28 1.55 Roundup 29 1.66 LH* (1 dimensional data) 29 1.6.11 LH* Addressing Scheme 30 1.6.22 LH* File Expansion 31 1.6.33 Conclusion 32 1.77 Orthogonal Aspects 33 1.7.11 Performance 33 1.7.22 Dimensions 33 1.7.33 Overhead 34 1.7.44 Distribution and Parallelism 34 1.7.55 Availability 35 22 The LH*LH Algorithm 37 2.11 Introduction 37 2.22 The Server 38 2.2.11 The LH Manager 38 2.2.22 LH* Partitioning of an LH File 39 2.2.33 Concurrent Request Processing and Splitting 42 2.2.44 Shipping 42 3 3 44 CONTENTS 2.33 Notes on LH*LH Communications 43 2.3.11 Communication Patterns 44 2.44 LH*LH Implementation 45 2.4.11 The System Initialization 45 2.4.22 The Data Client 46 2.4.33 The Server 50 2.4.44 Server Mapping 52 2.4.55 Summary and Future Work 54 2.4.66 Host for Scientific Data 55 2.55 Hardware Architecture 56 2.5.11 Communication 57 2.5.22 Measure Suite 58 2.66 Performance Evaluation 58 2.6.11 Scalability 59 2.6.22 Efficiency of Concurrent Splitting 64 2.77 Curiosity 68 2.88 Conclusion 69 33 SDDS for High-Performance Spatial Access 71 3.11 Introduction 72 3.22 hQT* Overview 73 3.2.11 Records 74 3.2.22 Pseudokey Construction 74 3.2.33 Bucket Numbering 75 3.2.44 Addressing 75 3.2.55 File Growth 77 3.33 Distribution in hQT* 77 3.3.11 Distribution (ForwardBuckets) 78 3.3.22 Distributed Point queries 78 3.3.33 Distributed Region Queries 78 3.3.44 IAM Policies 79 3.44 Server Splitting 80 3.4.11 hQT* Splitting 80 3.4.22 Dissection Splitting Algorithm 80 3.55 Measurements 81 3.5.11 Efficiency of IAM Policies 81 3.5.22 Server Load Distribution 85 3.5.33 Discussion 86 3.66 Conclusions 87 44 S7-storage: Multi-Attribute Storage 89 4.11 Introduction 90 4.22 Related Work 90 4.33 The O-storage 92 CONTENTSCONTENTS 5 4.3.11 Buckets and Branch nodes 92 4.3.22 An Example 93 4.3.33 Point Searching 94 4.3.44 Splitting Strategy 94 4.44 Performance evaluation and Tuning 95 4.4.11 Bucket Size vs Pruning 96 4.4.22 Insert costs 97 4.4.33 Search cost for a growing data set 98 4.4.44 Influence of Number of Attributes 100 4.4.55 Comparison with kd-tree 100 4.55 Exploration of the O-tree design space 101 4.5.11 Branches 101 4.5.22 Dynamic hash-function 103 4.5.33 Explored variants of fi-trees 103 4.5.44 Implementation Notes 104 4.66 Conclusions 105 III Applications of SDDSs 107 55 Database Systems 111 5.11 The Need for High Performance Databases Ill 5.22 Conventional Databases 113 5.33 Distributed Databases 113 5.44 Federated Databases 114 5.55 Multidatabases 114 5.66 Data Servers 115 5.77 Parallel Data Servers 115 5.88 Database Machines 117 5.99 Overview of Some Data Servers 118 5.100 DB History 120 5.111 Conclusions 121 5.122 Properties of Structures for Servers 122 5.12.11 The Problem 122 5.12.25.12.2 Scalability 122 5.12.33 Distribution 123 5.12.44 Availability 124 5.12.55 Conclusions 124 66 Scalable Distributed Storage Manager 125 6.11 Seamless SDDS integration in an Extensible DBMS 125 6.22 Introduction 126 6.33 Background 127 6.3.11 Scalable Distributed Data Structures 127 66 CONTENTS 6.3.22 Monet 128 6.44 SDDS within an Extensible Database System 128 6.4.11 SDDS requirements on a DBMS 128 6.4.22 Resource Management 129 6.4.33 SDDS Administration 129 6.4.44 Algebraic Operations 130 6.55 Implementation and Performance Study 131 6.5.11 Optimal Size of a Distributed Partition 132 6.5.22 Overhead added by SDDSs 134 6.5.33 Performance Scalability 134 6.5.44 Discussion 136 6.66 Summary 137 77 Summary & Future Issues 139 7.11 Summary 139 7.22 Extensions to this thesis 139 7.33 Future work 140 Listt of Figures 1.11 LH* File Expansion Scheme 30 2.11 The Data Server 39 2.22 The LH-structure 40 2.33 Pseudo-key usage by LH and LH* 40 2.44 Partitioning of an LH-file by LH* splitting 41 2.55 One node on the Parsytec machine 56 2.66 Static routing on a 64 nodes machine between two nodes. 57 2.77 Allocation of servers and clients 58 2.88 Build time of the file for a varying number of clients 59 2.99 Global insert time measure at one client, varying the number off clients 60 2.100 Actual throughput with varying number of clients 61 2.111 Ideal and actual throughput with respect to the number of clientss 62 2.122 Comparison between Static and Dynamic splitting strategy, onee client 63 2.133 Comparison between Static and Dynamic splitting, with four clientss 63 2.144 Efficiency of individual shipping 64 2.155 Efficiency of bulk shipping 65 2.166 Efficiency of the concurrent splitting 66 2.177 LH*LH client insert time scalability 67 3.11 The Record Structure 74 3.22 An offset space-filling curve first 3 layers 75 3.33 Navigation in a) Quad-Tree b) hQT* 76 3.44 Left: hQT* file key space partitioning by 4 successive splits. Right:: The equivalent quad-tree 77 3.55 Split Dissection Algorithm 81 3.66 a) Forward message count using different policies on servers 1,2,33 and 31.b) Only the 3 most efficient strategies 84 3.77 Split distribution over "time" 86 7 7 LISTLIST OF FIGURES 4.11 A bucket of an !7-tree and its attributes 92 4.22 A "typical" ft-marshaled tree 93 4.33 Bucket Split Algorithm 95 4.44 Varying a) search times b) insert times 97 4.55 Search time using a) 1,2,3,8 attributes in O-tree b) details of 88 attribute 98 4.66 a) Search times b) 8, 16 attributes files c) 8 attribute file e) 166 attribute file 99 4.77 8 attribute file, standard deviation for patterns search a) The kd-treee and fi-tree 10% compared with the 0-20% b) KD comparedd with fï-tree with acceptance limit of 10% and 20% 101 5.11 Data and application servers 116 6.11 Local memory on one node, varying sizes of data 133 6.22 One node, distributed access, varying sizes 134 6.33 Constant sized relation 8 MBytes, varying number of nodes.
Recommended publications
  • The Helios Operating System
    The Helios Operating System PERIHELION SOFTWARE LTD May 1991 COPYRIGHT This document Copyright c 1991, Perihelion Software Limited. All rights reserved. This document may not, in whole or in part be copied, photocopied, reproduced, translated, or reduced to any electronic medium or machine readable form without prior consent in writing from Perihelion Software Limited, The Maltings, Charlton Road, Shepton Mallet, Somerset BA4 5QE. UK. Printed in the UK. Acknowledgements The Helios Parallel Operating System was written by members of the He- lios group at Perihelion Software Limited (Paul Beskeen, Nick Clifton, Alan Cosslett, Craig Faasen, Nick Garnett, Tim King, Jon Powell, Alex Schuilen- burg, Martyn Tovey and Bart Veer), and was edited by Ian Davies. The Unix compatibility library described in chapter 5, Compatibility,im- plements functions which are largely compatible with the Posix standard in- terfaces. The library does not include the entire range of functions provided by the Posix standard, because some standard functions require memory man- agement or, for various reasons, cannot be implemented on a multi-processor system. The reader is therefore referred to IEEE Std 1003.1-1988, IEEE Stan- dard Portable Operating System Interface for Computer Environments, which is available from the IEEE Service Center, 445 Hoes Lane, P.O. Box 1331, Pis- cataway, NJ 08855-1331, USA. It can also be obtained by telephoning USA (201) 9811393. The Helios software is available for multi-processor systems hosted by a wide range of computer types. Information on how to obtain copies of the Helios software is available from Distributed Software Limited, The Maltings, Charlton Road, Shepton Mallet, Somerset BA4 5QE, UK (Telephone: 0749 344345).
    [Show full text]
  • ISRA VISION at a Glance
    ISRA VISION Initiating Coverage 18 February 2019 Better than the human eye – Growth story is expected continue In a highly fragmented market environment ISRA VISION ranks Target price (EUR) 36 Share price (EUR) 28 among the leading players for machine vision technologies with a strong focus on software solutions. Driven by market trends as well as R&D efforts, we expect ISRA’s growth story to continue and Forecast changes forecast a revenue growth CAGR of 10.4% until FY 2022/23e. EBIT % 2019e 2020e 2021e is expected to increase by 10.6% during this period. In addition to organic growth, M&A driven growth is very likely. We initiate the Revenues NM NM NM EBITDA NM NM NM coverage of ISRA VISION with Buy and a TP of EUR 36.00. EBIT adj NM NM NM EPS reported NM NM NM EPS adj NM NM NM ISRA VISION among the technology leaders Source: Pareto ISRA is one of the leading providers for machine vision solutions with specialisation in 3D machine vision. The company’s Ticker ISRG.DE, ISR GR competences are mainly concentrated around the internally Sector Industrials developed software application. Future demand is driven by more Shares fully diluted (m) 21.9 Market cap (EURm) 605 complexity and miniaturization, quality requirements and entering Net debt (EURm) -18 into new markets. Customer benefits are cost savings, process Minority interests (EURm) 2 optimization, higher flexibility and a ROI of clearly <1 year. Enterprise value 19e (EURm) 590 Free float (%) 65 Growth story should continue - 5yrs growth CAGR of 10.4% The well diversified customer base in various end-markets helps ISRA to reduce cyclicality.
    [Show full text]
  • STATUS REPORT Sc? (1@’Rq» Lu? Ul __ V? 211 — Q CERN Participation in the GP MIMD ESPRIT Project
    CERN LIBRARIES, GENEVA CERN/DRDU 9 1-12 Danc/M5 llltlllllllllulll II\)Illll|lI)I|l)H)|lllllI|ll)) afch , ;, — P A $(100000690 <_ fri · \ xg ·-J ’ STATUS REPORT Sc? (1@’rQ» lu? Ul __ V? 211 — Q CERN participation in the GP MIMD ESPRIT project P.C. Burkimsher, R.W. Dobinson ' , A. King, B. Martin, and I.M. Willers, CERN. J—L. Pages, University of Lausanne A. Schneider, University of Geneva. In collaboration with INMOS (UK), lead partner. Meiko (UK), Parsys (UK), Parsytec (D) and Telmat (F), main partners. and Grupo APD (E), INESC (P), Siemens (D), SICS (S), Southampton University (UK), IRISA (F) OCR Output spokesman z JMU Introduction A proposal was submitted to ESPRIT in January 1990 by INMOS and four European manufacturers of Transputer based computers. The aim of the proposal is to develop high performance parallel processing computers using a new generation of Transputers. ESPRIT is the EEC supported European Strategic Programme for Research and Development in Information Technology. The technical, manpower and budget details of the project are contained in the technical annex [l] of the official contract that has been signed by the lead and main project partners with the EEC. CERN has taken part in the preparatory work for the GP MIMD ESPRIT proposal and must now decide on the continuation of this work within the framework of the approved project. Transputer Technology Today's microprocessors make available on a single chip a significant fraction of the computing performance of a traditional mainframe. A recent trend has been to construct very powerful computing systems of interconnected microprocessors: up to hundreds or even thousands of processor nodes, each with local memory, joined by a fast switching network.
    [Show full text]
  • ALTERNATING-DIRECTION LINE-RELAXATION METHODS on MULTICOMPUTERS* J)RN HOFHAUS and ERIC E VAN DE VELDE$ Abstract
    SIAM J. ScI. COMPUT. () 1996 Society for Industrial and Applied Mathematics Vol. 17, No. 2, pp. 454-478, March 1996 010 ALTERNATING-DIRECTION LINE-RELAXATION METHODS ON MULTICOMPUTERS* J)RN HOFHAUS AND ERIC E VAN DE VELDE$ Abstract. We study the multicomputer performance of a three-dimensional Navier-Stokes solver based on alternating-direction line-relaxation methods. We compare several multicomputer implementations, each of which combines a particular line-relaxation method and a particular distributed block-tridiagonal solver. In our experiments, the problem size was determined by resolution requirements of the application. As a result, the granularity of the computations of our study is finer than is customary in the performance analysis of concurrent block-tridiagonal solvers. Our best results were obtained with a modified half-Gauss-Seidel line-relaxation method implemented by means of a new iterative block-tridiagonal solver that is developed here. Most computations were performed on the Intel Touchstone Delta, but we also used the Intel Paragon XP]S, the Parsytec SC-256, and the Fujitsu S-600 for comparison. Key words. Navier-Stokes equations, concurrency, parallelism, block-tridiagonal systems, tridiagonal systems, ADI, alternating directions AMS subject classifications. 65M20, 65N40, 65Y05, 76C05, 76M20 1. Introduction. When using alternating-direction line-relaxation methods for systems of partial-differential equations discretized on a rectangular grid, one must solve many block- tridiagonal systems of linear equations in every relaxation step. This type of computation surfaces in many applications. In our work, we faced it when parallelizing a highly vectorized solver for the three-dimensional, unsteady, and incompressible Navier-Stokes equations [4] for use on multicomputers.
    [Show full text]
  • Implementation of an Environment for Monte Carlo Simulation of Fully 3-D Positron Tomography on a High-Performance Parallel Platform
    Parallel Computing 24 (1998) 1523±1536 Implementation of an environment for Monte Carlo simulation of fully 3-D positron tomography on a high-performance parallel platform Habib Zaidi *, Claire Labbe, Christian Morel Division of Nuclear Medicine, Geneva University Hospital CH-1211 Geneva 4, Switzerland Received 15 January 1998; received in revised form 15 April 1998 Abstract 12/02/1998 ext-2001-007 This paper describes the implementation of the Eidolon Monte Carlo program designed to simulate fully three-dimensional (3-D) cylindrical positron tomographs on a MIMD parallel architecture. The original code was written in Objective-C and developed under the NeXT- STEP development environment. Dierent steps involved in porting the software on a parallel architecture based on PowerPC 604 processors running under AIX 4.1 are presented. Basic aspects and strategies of running Monte Carlo calculations on parallel computers are de- scribed. A linear decrease of the computing time was achieved with the number of computing nodes. The improved time performances resulting from parallelisation of the Monte Carlo calculations makes it an attractive tool for modelling photon transport in 3-D positron tomography. The parallelisation paradigm used in this work is independent from the chosen parallel architecture. Ó 1998 Elsevier Science B.V. All rights reserved. Keywords: Monte Carlo simulation; Positron emission tomography; Random numbers; Image recon- struction; Parallel computer 1. Introduction Positron emission tomography (PET) is a well-established imaging modality which provides physicians with information about the body's chemistry not available through any other procedure. Three-dimensional (3-D) PET provides qualitative and * Corresponding author. E-mail: [email protected] 0167-8191/98/$ ± see front matter Ó 1998 Elsevier Science B.V.
    [Show full text]
  • Logp: Towards a Realistic Model of Parallel Computation
    LogP: Towards a Realistic Model of Parallel Computation David Culler, Richard Karp,y David Patterson, Abhijit Sahay, Klaus Erik Schauser, Eunice Santos, Ramesh Subramonian, and Thorsten von Eicken Computer Science Division, University of California, Berkeley Abstract A vast body of theoretical research has focused either on overly simplistic models of parallel computation, notably the PRAM, or overly specific models that have few representatives in the real world. Both kinds of models encourage exploitation of formal loopholes, rather than rewarding development of techniques that yield performance across a range of current and future parallel machines. This paper offers a new parallel machine model, called LogP, that reflects the critical technology trends underlying parallel computers. It is intended to serve as a basis for developing fast, portable parallel algorithms and to offer guidelines to machine designers. Such a model must strike a balance between detail and simplicity in order to reveal important bottlenecks without making analysis of interesting problems intractable. The model is based on four parameters that specify abstractly the computing bandwidth, the communi- cation bandwidth, the communication delay, and the efficiency of coupling communication and computation. Portable parallel algorithms typically adapt to the machine configuration, in terms of these parameters. The utility of the model is demonstrated through examples that are implemented on the CM-5. Keywords: massively parallel processors, parallel models, complexity analysis, parallel algo- rithms, PRAM 1 Introduction Our goal is to develop a model of parallel computation that will serve as a basis for the design and analysis of fast, portable parallel algorithms, i.e., algorithms that can be implemented effectively on a wide variety of current and future parallel machines.
    [Show full text]
  • Uva-DARE (Digital Academic Repository)
    UvA-DARE (Digital Academic Repository) A Communication Kernel for Parallel Programming Support on a Massively Parallel Processor System Overeinder, B.J.; Vesseur, J.J.J.; van der Linden, F.; Sloot, P.M.A. Publication date 1995 Published in Proceedings of the Workshop on Parallel Programming and Computation (ZEUS'95) and the 4th Nordic Transputer Conference (NTUG'95) Link to publication Citation for published version (APA): Overeinder, B. J., Vesseur, J. J. J., van der Linden, F., & Sloot, P. M. A. (1995). A Communication Kernel for Parallel Programming Support on a Massively Parallel Processor System. In L. Finmo, & P. Fritzon (Eds.), Proceedings of the Workshop on Parallel Programming and Computation (ZEUS'95) and the 4th Nordic Transputer Conference (NTUG'95) (pp. 259-266). IOS Press. General rights It is not permitted to download or to forward/distribute the text or part of it without the consent of the author(s) and/or copyright holder(s), other than for strictly personal, individual use, unless the work is under an open content license (like Creative Commons). Disclaimer/Complaints regulations If you believe that digital publication of certain material infringes any of your rights or (privacy) interests, please let the Library know, stating your reasons. In case of a legitimate complaint, the Library will make the material inaccessible and/or remove it from the website. Please Ask the Library: https://uba.uva.nl/en/contact, or a letter to: Library of the University of Amsterdam, Secretariat, Singel 425, 1012 WP Amsterdam, The Netherlands. You will be contacted as soon as possible. UvA-DARE is a service provided by the library of the University of Amsterdam (https://dare.uva.nl) Download date:01 Oct 2021 A Communication Kernel for Parallel Programming Support on a Massively Parallel Processor System B.
    [Show full text]
  • TNMOC-Newsletter-Q1-2015.Pdf
    InSync Members’ News TNMoC will be hosting a Members’ Open Day on Saturday 28th March 2015, 10:00am for 10:30am. This is your opportunity to share your ideas regarding the future of the membership programme. Important changes are afoot both in the museum and in the Members’ organisation. We want you to be amongst the first to hear about them. In particular, we are keen to consult more with Members and to enable you to become more involved in the Museum. We’d like to share our plans with you during the open day and hear your views, so we hope that you will be able to attend. The programme for the day will include an exclusive, in-depth look at the Cambridge University EDSAC reconstruction. If you’ve been to the museum recently you’ll have seen that the project is progressing very well. Members of the EDSAC team will be there to tell you all about it. Volunteers and the Museum’s management team will be on hand to chat with Members about the museum and its ever-changing range of exhibits. If there is sufficient interest we will arrange one or more specialist guided tours for Members. Full details will be sent to you nearer the time. If you have any questions you are invited to contact John Linford at [email protected]. In the meantime, please put the Open Day in your diary. We look forward to seeing you! Connecting with the Museum of Computing Keep up to date with the latest information from the museum by visiting our various social media and website presences.
    [Show full text]
  • A High-Performance, Portable Implementation of the MPI Message
    A HighPerformance Portable Implementation of the MPI Message Passing Interface Standard William Gropp Ewing Lusk Mathematics and Computer Science Division Argonne National Lab oratory Nathan Doss Anthony Skjellum Department of Computer Science NSF Engineering Research Center for CFS y Mississippi State University Abstract MPI Message Passing Interface is a sp ecication for a standard library for message passing that was dened by the MPI Forum a broadly based group of parallel com puter vendors library writers and applications sp ecialists Multiple implementations of MPI havebeendevelop ed In this pap er we describ e MPICH unique among existing implementations in its design goal of combining p ortability with high p erformance We do cument its p ortability and p erformance and describ e the architecture bywhichthese features are simultaneously achieved We also discuss the set of to ols that accompany the free distribution of MPICH which constitute the b eginnings of a p ortable parallel programming environment A pro ject of this scop e inevitably imparts lessons ab out parallel computing the sp ecication b eing followed the currenthardware and software environment for parallel computing and pro ject management we describ e those we have learned Finallywe discuss future developments for MPICH including those nec essary to accommo date extensions to the MPI Standard now b eing contemplated by the MPI Forum Intro duction The messagepassing mo del of parallel computation has emerged as an expressive e cient and wellundersto o d paradigm
    [Show full text]
  • The Helios File Server Installation Information
    The Helios File Server Installation Information PERIHELION SOFTWARE LTD November 1990 Copyright This document Copyright °c 1991, Perihelion Software Limited. All rights re- served. This document may not, in whole or in part be copied, photocopied, reproduced, translated, or reduced to any electronic medium or machine read- able form without prior consent in writing from Perihelion Software Limited, The Maltings, Charlton Road, Shepton Mallet, Somerset BA4 5QE, UK. This manual is Edition 2.0, November 1990. Acorn and ARM are trademarks of Acorn Computers Ltd. Amiga is a registered trademark of Commodore-Amiga, Inc. Apple is a registered trademark of Apple Computers, Inc. Atari is a trademark of the Atari Corporation. Commodore is a registered trademark of Commodore Electronics, Ltd. Ethernet is a trademark of Xerox Corporation. Helios is a trademark of Perihelion Software Limited. IBM is a registered trademark of International Business Machines, Inc. Inmos, occam, T414, T425 and T800 are trademarks of the Inmos group of companies. Intel and iPSC are registered trademarks of Intel Corporation. Macintosh is a trademark of Apple Computers, Inc. Meiko and Cesius are trademarks of Meiko Limited. Motorola is a trademark of Motorola, Inc. MS-DOS is a registered trademark of The Microsoft Corporation. Parsytec, Paracom and SuperCluster are trademarks of Parsytec GmbH. POSIX refers to the standard de¯ned by IEEE Standard 1003.1-1988; Posix refers to the library calls based upon this standard. Transtech is a trademark of Transtech Devices Ltd. Sun, SunOs and SunView are trademarks of Sun Microsystems. Telmat and T.Node are trademarks of Telmat Informatique. Unix is a registered trademark of AT&T.
    [Show full text]
  • Versatility of Bulk Synchronous Parallel Computing: from the Heterogeneous Cluster to the System on Chip
    Dissertation Versatility of Bulk Synchronous Parallel Computing: From the Heterogeneous Cluster to the System on Chip Olaf Bonorden Fakultät für Elektrotechnik, Informatik und Mathematik Institut für Informatik und Heinz Nixdorf Institut February 2008 ii Zusammenfassung Der Bedarf an Rechenleistung in den Wissenschaften und der Industrie steigt ständig an. Bei der Entwicklung schnellerer Prozessoren gelangt man allerdings an physi- kalische Grenzen, so dass zur weiteren Leistungssteigerung sehr viele Prozessoren zusammen als Parallelcomputer benutzt werden müssen. Zur Nutzung dieser Super- computer ist es wichtig, allgemeine Modelle und Werkzeuge zu haben, die es erlau- ben, unabhängig von der speziellen Hardware effiziente Algorithmen zu entwickeln und zu implementieren. In dieser Dissertation werden Modelle für parallele Systeme vorgestellt, ein Über- blick über Algorithmen für diese Modelle gegeben und effiziente Implementierungen für verschiedene Architekturen entwickelt. Der Schwerpunkt liegt dabei auf der Fami- lie der Bulk Synchronous Parallel Modelle, da diese die Entwicklung portabler, aber trotzdem effizienter paralleler Programme erlauben. Für die Implementierungen werden zwei Architekturen betrachtet: ein On-Chip- Parallelcomputer und Workstation-Cluster. Im Bereich des On-Chip-Systems zeigt die Arbeit, wie das benutzte Modell die Entwicklung applikationsunabhängiger, effizien- ter, paralleler Systeme unterstützen kann. Die Workstation-Cluster, auf denen nur freie Rechenkapazitäten genutzt werden dürfen, stehen auf der anderen Seite des Spektrums paralleler Systeme. Sie unter- scheiden sich vom On-Chip-System nicht nur durch viel größere Latenzen, geringe- re Kommunikationsbandbreite und größeren Arbeitsspeicher, sie können auch hete- rogen sein, d. h., verschiedene Computertypen enthalten. Hierdurch und durch die variable, sich ständig ändernde, nutzbare Rechenkapazität der einzelnen Knoten er- geben sich besondere Herausforderungen, z. B.
    [Show full text]
  • A Parallel Implementation of Genetic Programming That Achieves Super-Linear Performance
    A PARALLEL IMPLEMENTATION OF GENETIC PROGRAMMING THAT ACHIEVES SUPER-LINEAR PERFORMANCE David Andre John R. Koza Visiting Scholar, Computer Science Computer Science Department Department Stanford University Stanford University EMAIL: [email protected] EMAIL: [email protected] WWW: http://www-cs- WWW: http://www- faculty.stanford.edu/~koza/ leland.stanford.edu/~phred/ Abstract: This paper describes the successful parallel implementation of genetic programming on a network of processing nodes using the transputer architecture. With this approach, researchers of genetic algorithms and genetic programming can acquire computing power that is intermediate between the power of currently available workstations and that of supercomputers at intermediate cost. This approach is illustrated by a comparison of the computational effort required to solve a benchmark problem. Because of the decoupled character of genetic programming, our approach achieved a nearly linear speed up from parallelization. In addition, for the best choice of parameters tested, the use of subpopulations delivered a super linear speed-up in terms of the ability of the algorithm to solve the problem. Several examples are also presented where the parallel genetic programming system evolved solutions that are competitive with human performance on the same problem. 1 Introduction Amenability to parallelization is an appealing feature of genetic algorithms, evolutionary programming, evolution strategies, classifier systems, and genetic programming [1][2][3][4]. The probability of success in applying the genetic algorithm to a particular problem often depends on the adequacy of the size of the population in relation to the difficulty of the problem. Although many moderately sized problems can be solved by the genetic algorithm using currently available single-processor workstations, more substantial problems usually require larger populations.
    [Show full text]