Distributed, File Based, Geospatial Array DBMS

Total Page:16

File Type:pdf, Size:1020Kb

Distributed, File Based, Geospatial Array DBMS ChronosDB: Distributed, File Based, Geospatial Array DBMS Ramon Antonio Rodriges Zalipynis National Research University Higher School of Economics, Moscow, Russia [email protected] ABSTRACT that are very accustomed to them. NCO (NetCDF Oper- An array DBMS streamlines large N-d array management. ators) have been under development since about 1995 [43], A large portion of such arrays originates from the geospatial GDAL (Geospatial Data Abstraction Library) provides tools for managing over 150 raster file formats, has 106 lines of domain. The arrays often natively come as raster files while ≈ standalone command line tools are one of the most popu- code and hundreds of contributors [14]. Many tools utilize lar ways for processing these files. Decades of development multicore CPUs but mostly work on a single machine. and feedback resulted in numerous feature-rich, elaborate, The in situ technique has gained increased attention due free and quality-assured tools optimized mostly for a single to the explosive growth of raster data volumes in diverse file machine. ChronosDB partially delegates in situ data pro- formats [3, 85, 6, 8]. Unlike the in-db approach, the in situ cessing to such tools and offers a formal N-d array data approach operates on data in their original file formats in a model to abstract from the files and the tools. ChronosDB standard file system and does not require importing into an readily provides a rich collection of array operations at scale internal DBMS format. Files may be preprocessed before and outperforms SciDB by up to 75 on average. querying but this is usually much faster than a full import. × Keeping data of an array DBMS in open, widely adopted, PVLDB Reference Format: and standardized file formats has numerous advantages: eas- R.A. Rodriges Zalipynis. ChronosDB: Distributed, File Based, ier data sharing, powerful storage capabilities, the absence Geospatial Array DBMS. PVLDB, 11(10): 1247-1261, 2018. of a time-consuming and error-prone import phase for many DOI: https://doi.org/10.14778/3231751.3231754 data types, direct accessibility of DBMS data to other sys- tems, and easier migration to other DBMS to name a few [59]. 1. INTRODUCTION Many raster processing algorithms require significant im- A georeferenced raster or more generally an N-d array is plementation efforts, but already existing stable and multi- the primary data type in a broad range of fields including functional tools are largely ignored in this research trend. climatology, geology, and remote sensing which are expe- Algorithms are re-implemented almost from scratch delay- riencing tremendous growth of data volumes. For example, ing the emergence of an array DBMS being competitive with DigitalGlobe is a commercial satellite imagery provider that existing operational pipelines, e.g. SciDB appeared in 2008 collects 70 TB of imagery on an average day [35]. and still lacks even core operations like interpolation [65]. Traditionally rasters are stored in files, not in databases. The idea of partially delegating in situ array operations The European Centre for Medium-Range Weather Forecasts to existing tools was first realized in ChronosServer [59, 58, (ECMWF) has alone accumulated 137:5 106 files sized 11, 60, 62, 61]. Its successor is ChronosDB extended with 52.7 PB in total [26]. This file-centric model× resulted in a a significantly improved data model (section 2), new com- broad set of sophisticated raster file formats. For example, ponents, optimizations, array management, query execution GeoTIFF is an effort by over 160 different remote sensing, capabilities (section 3), and array operations (section 4). GIS (Geographic Information System), cartographic, and The ChronosDB formal data model uniformly represents surveying related companies and organizations [23]. NetCDF diverse raster data types and formats, takes into account the has been under development since 1990 [56], is standardized distributed context, and is independent of the underlying by OGC [44], and supports multidimensional arrays, chunk- raster file formats and tools at the same time (section 2). ing, compression, and hierarchical name space [40]. ChronosDB distributed algorithms are based on the model, Command line tools have long being developed to man- unified, formalized, and targeted at the in situ processing of age raster files. Many tools have large user communities arbitrary geospatial N-d arrays (section 4). The algorithms are designed to always delegate signifi- Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are cant portions of their work to the tools. This proves wide not made or distributed for profit or commercial advantage and that copies applicability of the tools and the delegation approach. The bear this notice and the full citation on the first page. To copy otherwise, to delegation happens by direct submission of files to a tool as republish, to post on servers or to redistribute to lists, requires prior specific command line arguments to process them on a single cluster permission and/or a fee. Articles from this volume were invited to present node. ChronosDB re-partitions and streams input/output their results at The 44th International Conference on Very Large Data Bases, files between the nodes and tools to scale out the processing. August 2018, Rio de Janeiro, Brazil. Proceedings of the VLDB Endowment, Vol. 11, No. 10 In summary, the major contributions of our work include: Copyright 2018 VLDB Endowment 2150-8097/18/06... $ 10.00. (1) we show that it is possible to build a well-abstracted, DOI: https://doi.org/10.14778/3231751.3231754 1247 extensible, and efficient distributed geospatial N-d array a totally ordered set, di[j] < di[j + 1], and di[j] = NA for 6 DBMS by leveraging existing elaborate command line tools j Di. In this case, A is defined as A(d1; d2; : : : ; dN ): T. 8 2 designed for a single machine; A hyperslab A0 A is an N-d subarray of A. The hyper- v (2) we give a holistic description of the complete DBMS slab A0 is defined by the notation A[b1 : e1; : : : ; bN : eN ] = leveraging the tools: ChronosDB has a relatively simple yet A0(d10 ; : : : ; dN0 ), where bi; ei Z, 0 bi ei < li, di0 = 2 6 6 flexible coordination layer on top of powerful components. di[bi : ei], d0 = ei bi + 1, and for all yi [0; ei bi] the j ij − 2 − This enables exceptional performance, rich functionality and following holds: A0[y1; : : : ; yN ] = A[y1 + b1; : : : ; yN + bN ], avoids re-implementing array algorithms from scratch. It is di0 [yi] = di[yi + bi](A and A0 have a common coordinate also easier to manage the development life cycle of such a subspace over which cell values of A and A0 coincide). The DBMS as it is comprehensive by a small team of engineers; dimensionality of A and A0 is the same. We will omit \: ei" (3) we thoroughly compare ChronosDB and SciDB on up if bi = ei or \bi : ei" if bi = 0 and ei = di 1. to 32-node clusters in the Cloud on real-world data: Landsat Arrays X and Y overlap iff Q : Q j Xj − Q Y . Array satellite scenes and climate reanalysis. We took SciDB as it Q is called the greatest common9 hyperslabv ^of Xvand Y and is the only freely available distributed array DBMS to date. denoted by gch(X; Y ) iff @W :(W X) (W Y ) (Q W ) (Q = W ). An array X coversv an array^ Yviff Y ^ X.v ^ 6 v 2. CHRONOSDB DATA MODEL 2.3 Datasets Two dataset types exist: raw and regular. Raw datasets 2.1 Motivation capture a broad range of possible raster data types, e.g. a The most widely used industry-standard data models for set of scattered, overlapping satellite scenes, fig. 1. A raw abstracting from raster file formats are Unidata CDM, GDAL dataset must be \cooked" into a regular one in order to Data Model and ISO 19123. Industrial models are map- perform array operations on it, section 3.3. Many real-world pable to each other [40] and have resulted from decades of array data already satisfy regular dataset criteria and need considerable practical experience. However, they work with not to be cooked, e.g. gridded geophysical data, section 5. a single file, not with a set of files as a single array. Both datasets are two-level and contain a user-level array The most well-known research models and algebras for and a set of system-level arrays (array and subarrays for dense N-d, general-purpose arrays are AML [38], AQL [37], short). Subarrays are distributed among cluster nodes and and RAM [81]. They are mappable to Array Algebra [4]. stored as ordinary raster files. An array is never materialized The creation of the ChronosDB data model was motivated and stored explicitly: an operation with an array is mapped by the following features that are not simultaneously present to a sequence of operations with respective subarrays. in the existing data models: (1) the treatment of arrays in Datasets are read-only: an array operation produces a multiple files arbitrarily distributed over cluster nodes as a new dataset. This has a strong practical motivation. The single array, (2) formalized industrial experience to leverage majority of raster data come from instrumental observation it in the algorithms, (3) a rich set of data types (Gaussian, and numerical simulation. Original data are never changed irregular grids, etc.), (4) subarray mapping to a raster file is once they are produced. Derivative data differ vastly: raster almost 1:1 but still independent from a format. As can be algorithms usually alter large portions of an array. raw raw seen from [40], the N-d array model (section 2.2) resembles Formally, a raw dataset D = (A; M; P ) has a user- CDM while the two-level set-oriented dataset model (sec- level array A(d1; d2; : : : ; dN ): T, metadata M, and a set raw tion 2.3) provides additional necessary abstractions.
Recommended publications
  • Array DBMS in Environmental Science
    The 9th IEEE International Conference on Intelligent Data Acquisition and Advanced Computing Systems: Technology and Applications 21-23 September, 2017, Bucharest, Romania Array DBMS in Environmental Science: Satellite Sea Surface Height Data in the Cloud Ramon Antonio Rodriges Zalipynis National Research University Higher School of Economics, Moscow, Russia [email protected] Abstract – Nowadays environmental science experiences data model is designed to uniformly represent diverse tremendous growth of raster data: N-dimensional (N-d) raster data types and formats, take into account a dis- arrays coming mainly from numeric simulation and Earth tributed cluster environment, and be independent of the remote sensing. An array DBMS is a tool to streamline raster data processing. However, raster data are usually stored in underlying raster file formats at the same time. Also, new files, not in databases. Moreover, numerous command line distributed algorithms are proposed based on the model. tools exist for processing raster files. This paper describes a Four modern raster data management trends are rele- distributed array DBMS under development that partially vant to this paper: industrial raster data models, formal delegates raster data processing to such tools. Our DBMS offers a new N-d array data model to abstract from the array models and algebras, in situ data processing algo- files and the tools and processes data in a distributed fashion rithms, and raster (array) DBMS. Good survey on the directly in their native file formats. As a case study, popular algorithms is contained in [7]. A recent survey of existing satellite altimetry data were used for the experiments carried array DBMS and similar systems is in [5].
    [Show full text]
  • An Array DBMS for Simulation Analysis and ML Models Predictions
    SAVIME: An Array DBMS for Simulation Analysis and ML Models Predictions Hermano. L. S. Lustosa1, Anderson C. Silva1, Daniel N. R. da Silva1, Patrick Valduriez2, Fabio Porto1 1 National Laboratory for Scientific Computing, Rio de Janeiro, Brazil {hermano, achaves, dramos, fporto}@lncc.br 2 Inria, University of Montpellier, CNRS, LIRMM, France [email protected] Abstract. Limitations in current DBMSs prevent their wide adoption in scientific applications. In order to make them benefit from DBMS support, enabling declarative data analysis and visualization over scientific data, we present an in-memory array DBMS called SAVIME. In this work we describe the system SAVIME, along with its data model. Our preliminary evaluation show how SAVIME, by using a simple storage definition language (SDL) can outperform the state-of-the-art array database system, SciDB, during the process of data ingestion. We also show that it is possible to use SAVIME as a storage alternative for a numerical solver without affecting its scalability, making it useful for modern ML based applications. Categories and Subject Descriptors: H.2.4 [Database Management]: Systems Keywords: Scientific Data Management, Multidimensional Array, Machine Learning 1. INTRODUCTION Due to the increasing computational power of HPC environments, vast amounts of data are now generated in different research fields, and a relevant part of this data is best represented as array data. Geospatial and temporal data in climate modeling, astronomical images, medical imagery, multimedia and simulation data are naturally represented as multidimensional arrays. To analyze such data, DBMSs offer many advantages, like query languages, which eases data analysis and avoids the need for extensive coding/scripting, and a logical data view that isolates data from the applications that consume it.
    [Show full text]
  • Array Databases: Concepts, Standards, Implementations
    Baumann et al. J Big Data (2021) 8:28 https://doi.org/10.1186/s40537-020-00399-2 SURVEY PAPER Open Access Array databases: concepts, standards, implementations Peter Baumann , Dimitar Misev, Vlad Merticariu and Bang Pham Huu* *Correspondence: b. Abstract phamhuu@jacobs-university. Multi-dimensional arrays (also known as raster data or gridded data) play a key role in de Large-Scale Scientifc many, if not all science and engineering domains where they typically represent spatio- Information Systems temporal sensor, image, simulation output, or statistics “datacubes”. As classic database Research Group, Jacobs technology does not support arrays adequately, such data today are maintained University, Bremen, Germany mostly in silo solutions, with architectures that tend to erode and not keep up with the increasing requirements on performance and service quality. Array Database systems attempt to close this gap by providing declarative query support for fexible ad-hoc analytics on large n-D arrays, similar to what SQL ofers on set-oriented data, XQuery on hierarchical data, and SPARQL and CIPHER on graph data. Today, Petascale Array Database installations exist, employing massive parallelism and distributed processing. Hence, questions arise about technology and standards available, usability, and overall maturity. Several papers have compared models and formalisms, and benchmarks have been undertaken as well, typically comparing two systems against each other. While each of these represent valuable research to the best of our knowledge there is no comprehensive survey combining model, query language, architecture, and practical usability, and performance aspects. The size of this comparison diferentiates our study as well with 19 systems compared, four benchmarked to an extent and depth clearly exceeding previous papers in the feld; for example, subsetting tests were designed in a way that systems cannot be tuned to specifcally these queries.
    [Show full text]
  • The Gamma Operator for Big Data Summarization on an Array DBMS
    JMLR: Workshop and Conference Proceedings 36:88{103, 2014 BIGMINE 2014 The Gamma Operator for Big Data Summarization on an Array DBMS Carlos Ordonez [email protected] Yiqun Zhang [email protected] Wellington Cabrera [email protected] University of Houston, USA Editors: Wei Fan, Albert Bifet, Qiang Yang and Philip Yu Abstract SciDB is a parallel array DBMS that provides multidimensional arrays, a query language and basic ACID properties. In this paper, we introduce a summarization matrix operator that computes sufficient statistics in one pass and in parallel on an array DBMS. Such sufficient statistics benefit a big family of statistical and machine learning models, including PCA, linear regression and variable selection. Experimental evaluation on a parallel cluster shows our matrix operator exhibits linear time complexity and linear speedup. Moreover, our operator is shown to be an order of magnitude faster than SciDB built-in operators, two orders of magnitude faster than SQL queries on a fast column DBMS and even faster than the R package when the data set fits in RAM. We show SciDB operators and the R package fail due to RAM limitations, whereas our operator does not. We also show PCA and linear regression computation is reduced to a few minutes for large data sets. On the other hand, a Gibbs sampler for variable selection can iterate much faster in the array DBMS than in R, exploiting the summarization matrix. Keywords: Array, Matrix, Linear Models, Summarization, Parallel 1. Introduction Row DBMSs remain the best technology for OLTP [Stonebraker et al.(2007)] and col- umn DBMSs are becoming a competitor in OLAP (Business Intelligence) query processing [Stonebraker et al.(2005)] in large data warehouses [Stonebraker et al.(2010)].
    [Show full text]
  • Multidimensional Arrays for Analysing Geoscientific Data
    International Journal of Geo-Information Review Multidimensional Arrays for Analysing Geoscientific Data Meng Lu 1,2,*,† , Marius Appel 1 and Edzer Pebesma 1 1 Institute for Geoinformatics, University of Muenster, 48149 Muenster, Germany; [email protected] (M.A.); [email protected] (E.P.) 2 Department of Physical Geography, Faculty of Geoscience, Utrecht University, 3584 CB Utrecht, The Netherlands * Correspondence: [email protected]; Tel.: +31 648901629 † Current address: Vening Meinesz Building A, Princetonlaan 8A, 3584 CB Utrecht, The Netherlands. Received: 27 June 2018; Accepted: 30 July 2018; Published: 3 August 2018 Abstract: Geographic data is growing in size and variety, which calls for big data management tools and analysis methods. To efficiently integrate information from high dimensional data, this paper explicitly proposes array-based modeling. A large portion of Earth observations and model simulations are naturally arrays once digitalized. This paper discusses the challenges in using arrays such as the discretization of continuous spatiotemporal phenomena, irregular dimensions, regridding, high-dimensional data analysis, and large-scale data management. We define categories and applications of typical array operations, compare their implementation in open-source software, and demonstrate dimension reduction and array regridding in study cases using Landsat and MODIS imagery. It turns out that arrays are a convenient data structure for representing and analysing many spatiotemporal phenomena. Although the array model simplifies data organization, array properties like the meaning of grid cell values are rarely being made explicit in practice. Keywords: multidimensional arrays; geoscientific data; data analysis; spatiotemporal modeling 1. Introduction An array is a storage form for a sequence of objects of similar type.
    [Show full text]
  • A Survey on Array Storage, Query Languages, and Systems
    A Survey on Array Storage, Query Languages, and Systems Florin Rusu Yu Cheng University of California, Merced 5200 N Lake Road Merced, CA 95343 {frusu,ycheng4}@ucmerced.edu February 2013 Abstract Since scientific investigation is one of the most important providers of massive amounts of ordered data, there is a renewed interest in array data processing in the context of Big Data. To the best of our knowledge, a unified resource that summarizes and analyzes array processing research over its long existence is currently missing. In this survey, we provide a guide for past, present, and future research in array processing. The survey is organizedalong three main topics. Array storage discusses all the aspects related to array partitioning into chunks. The identification of a reduced set of array operators to form the foundation for an array query language is analyzed across multiple such proposals. Lastly, we survey real systems for array processing. The result is a thorough survey on array data storage and processing that should be consulted by anyone interested in this research topic, independent of experience level. The survey is not complete though. We greatly appreciate pointers towards any work we might have forgotten to mention. 1 Introduction Big Data [33] is the new buzz word in computer science as of 2012. There are new conferences orga- nized specifically to tackle Big Data issues. Many classical research areas – beyond data management and databases – allocate significant attention to Big Data problems. And, most importantly, state governments provide unprecedented amounts of funds to support Big Data research [36]—at the end of the day, Big Data analytics played a significant role in the 2012 US presidential elections.
    [Show full text]
  • The Bigdawg Polystore System
    The BigDAWG Polystore System Jennie Duggan Aaron J. Elmore Michael Magda Balazinska Bill Howe Northwestern Univ. of Chicago Stonebraker Univ. of Univ. of MIT Washington Washington Jeremy Kepner Sam Madden David Maier Tim Mattson Stan Zdonik MIT MIT Portland State Intel Brown Univ. ABSTRACT and Technology Center for Big Data, we have constructed This paper presents a new view of federated databases to a medical example of this new class of applications, based address the growing need for managing information that on the MIMIC II dataset [18]. These publicly available pa- spans multiple data models. This trend is fueled by the tient records cover 26,000 intensive care unit admissions proliferation of storage engines and query languages based at Boston’s Beth Israel Deaconess Hospital. It includes on the observation that “no one size fits all”. To address waveform data (up to 125 Hz measurements from bed- this shift, we propose a polystore architecture; it is de- side devices), patient metadata (name, age, etc.), doctor’s signed to unify querying over multiple data models. We and nurse’s notes (text), lab results, and prescriptions filled consider the challenges and opportunities associated with (semi-structured data). A production implementation would polystores. Open questions in this space revolve around store all of the historical data augmented by real-time streams query optimization and the assignment of objects to stor- from current patients. Given the variety of data sources, age engines. We introduce our approach to these topics this system must support an assortment of data types, stan- and discuss our prototype in the context of the Intel Sci- dard SQL analytics (e.g., how many patients were given ence and Technology Center for Big Data.
    [Show full text]
  • The Architecture of Scidb
    The Architecture of SciDB Michael Stonebraker, Paul Brown, Alex Poliakov, Suchi Raman Paradigm4, Inc. 186 Third Avenue Waltham, MA 02451 Abstract. SciDB is an open-source analytical database oriented toward the data management needs of scientists. As such it mixes statistical and linear algebra operations with data management ones, using a natural nested multi-dimensional array data model. We have been working on the code for two years, most recently with the help of venture capital backing. Release 11.06 (June 2011) is downloadable from our website (SciDB.org). This paper presents the main design decisions of SciDB. It focuses on our decisions concerning a high-level, SQL-like query language, the issues facing our query optimizer and executor and efficient storage management for arrays. The paper also discusses implementation of features not usually present in DBMSs, including version control, uncertainty and provenance. Keywords: scientific data management, multi-dimensional array, statistics, linear algebra 1 Introduction and Background The Large Synoptic Survey Telescope (LSST) [1] is the next “big science” astronomy project, a telescope being erected in Chile, which will ultimately collect and manage some 100 Petabytes of raw and derived data. In October 2007, the members of the LSST data management team realized the scope of their data management problem, and that they were uncertain how to move forward. As a result, they organized the first Extremely Large Data Base (XLDB-1) conference at the Stanford National Accelerator Laboratory [2]. Present were many scientists from a variety of natural science disciplines as well as representatives from large web properties. All reported the following requirements: Multi-petabyte amounts of data.
    [Show full text]
  • On the Integration of Array and Relational Models in Databases
    On the Integration of Array and Relational Models in Databases by Dimitar Miˇsev A thesis submitted in partial fulfillment for the degree of Doctor of Philosophy in Computer Science Prof. Dr. Peter Baumann Jacobs University Bremen Prof. Dr. Michael Sedlmair Jacobs University Bremen Prof. Dr. Tore Risch Uppsala University Dr. Heinrich Stamerjohanns Jacobs University Bremen Date of defense: May 15th, 2018 Computer Science & Electrical Engineering Statutory Declaration Family Name, Given/First Name Miˇsev,Dimitar Matriculation number 20327580 Type of thesis PhD English: Declaration of Authorship I hereby declare that the thesis submitted was created and written solely by myself without any external support. Any sources, direct or indirect, are marked as such. I am aware of the fact that the contents of the thesis in digital form may be revised with regard to usage of unauthorized aid as well as whether the whole or parts of it may be identified as plagiarism. I do agree my work to be entered into a database for it to be compared with existing sources, where it will remain in order to enable further comparisons with future theses. This does not grant any rights of reproduction and usage, however. This document was neither presented to any other examination board nor has it been published. German: Erkl¨arung der Autorenschaft (Urheberschaft) Ich erkl¨arehiermit, dass die vorliegende Arbeit ohne fremde Hilfe ausschließlich von mir erstellt und geschrieben worden ist. Jedwede verwendeten Quellen, direkter oder indirekter Art, sind als solche kenntlich gemacht worden. Mir ist die Tatsache bewusst, dass der Inhalt der Thesis in digitaler Form gepr¨uftwerden kann im Hinblick darauf, ob es sich ganz oder in Teilen um ein Plagiat handelt.
    [Show full text]
  • Scidb DBMS Research at M.I.T
    SciDB DBMS Research at M.I.T. Michael Stonebraker, Jennie Duggan, Leilani Battle, Olga Papaemmanouil fstonebraker, jennie, [email protected], [email protected] Abstract This paper presents a snapshot of some of our scientific DBMS research at M.I.T. as part of the Intel Science and Technology Center on Big Data. We focus our efforts primarily on SciDB, although some of our work can be used for any backend DBMS. We summarize our work on making SciDB elastic, providing skew-aware join strategies, and producing scalable visualizations of scientific data. 1 Introduction In [19] we presented a description of SciDB, an array-based parallel DBMS oriented toward science applications. In that paper we described the tenets on which the system is constructed, the early use cases where it has found acceptance, and the state of the software at the time of publication. In this paper, we consider a collection of research topics that we are investigating at M.I.T. as part of the Intel Science and Technology Center on Big Data [20]. We begin in Section 2 with the salient characteristics of science data that guide our explorations. We then consider algorithms for making a science DBMS elastic, a topic we cover in Section 3. Then, we turn in Section 4 to query processing algorithms appropriate for science DBMS applications. Lastly, in Section 5 we discuss our work on producing a scalable visualization system for science applications. 2 Characteristics of Science DBMS Applications In this section we detail some of the characteristics of science applications that guide our explorations, specifi- cally an array data model, variable density of data, skew, and the need for visualization.
    [Show full text]
  • Principles of Distributed Database Systems: Spotlight on Newsql
    Principles of Distributed Database Systems: spotlight on NewSQL Patrick Valduriez Outline • Distributed database systems • NoSQL • Polystores • Spotlight on NewSQL • Taxonomy of NewSQL systems • Current trends Principles of Distributed Database Systems Tamer Özsu & Patrick Valduriez The Story of the Book Relational databases. 1991 “In the following 10 years, centralized DBMSs would be an antique curiosity and most organizations would move towards distributed DBMSs.” M. Stonebraker (1988) Advanced transaction models, 1999 query optimization, object data management, parallel DBMSs. 2011 Data replication, database clusters, web data integration, P2P, cloud. 2020 Blockchain, big data, data streaming, graph data analytics, NoSQL, NewSQL, polystores. SBBD 2020 © P. Valduriez 2020 4 Distributed Database - User View 1991 SBBD 2020 © P. Valduriez 2020 5 Distributed Database - User View 2020 SBBD 2020 © P. Valduriez 2020 6 Distributed DBMS – Reality User Query DBMS User Application Software DBMS Software DBMS Communication Software Subsystem User DBMS User Application Software Query DBMS Software User Query SBBD 2020 © P. Valduriez 2020 7 Definitions • A distributed database (DDB) is a collection of multiple, logically interrelated databases distributed over a computer network • WAN, LAN, cluster interconnect • A distributed database system (distributed DBMS) is the software that manages the DDB and provides an access mechanism that makes this distribution transparent to the users SBBD 2020 © P. Valduriez 2020 8 Promises of Distributed DBMSs • Transparent management of distributed, fragmented, and replicated data • Declarative query language (SQL) • Improved reliability/availability through replication and distributed transactions • Strong ACID consistency • Failover and online recovery • Improved performance • Proximity of data to its points of use • Data-based parallelism • Query optimization • Easier and more economical system expansion • Elasticity in the cloud SBBD 2020 © P.
    [Show full text]
  • Skew-Aware Join Optimization for Array Databases
    Skew-Aware Join Optimization for Array Databases Jennie Duggan‡, Olga Papaemmanouil†, Leilani Battle⋆, Michael Stonebraker⋆ ‡ Northwestern University, † Brandeis University, ⋆ MIT ‡ [email protected], †[email protected], ⋆{leilani, stonebraker}@csail.mit.edu ABSTRACT In addition, queries to scientific databases do not resemble the Science applications are accumulating an ever-increasing amount ones found in traditional business data processing applications. of multidimensional data. Although some of it can be processed in Complex analytics, such as predictive modeling and linear regres- a relational database, much of it is better suited to array-based en- sion, are prevalent in such workloads, replacing the more traditional gines. As such, it is important to optimize the query processing of SQL aggregates found in business intelligence applications. Such these systems. This paper focuses on efficient query processing of analytics are invariably linear algebra-based and are more CPU- join operations within an array database. These engines invariably intensive than RDBMS ones. “chunk” their data into multidimensional tiles that they use to effi- The relational data model has shown itself to be ill-suited for ciently process spatial queries. As such, traditional relational algo- many of these science workloads [37], and performance can be rithms need to be substantially modified to take advantage of array greatly improved through use of a multidimensional, array-based tiles. Moreover, most n-dimensional science data is unevenly dis- data model [34]. As a result, numerous array processing solutions tributed in array space because its underlying observations rarely have emerged to support science applications in a distributed envi- follow a uniform pattern.
    [Show full text]