The SQL++ Unifying Semi-Structured Query Language, and an Expressiveness Benchmark of SQL-On-Hadoop, Nosql and Newsql Databases

Total Page:16

File Type:pdf, Size:1020Kb

The SQL++ Unifying Semi-Structured Query Language, and an Expressiveness Benchmark of SQL-On-Hadoop, Nosql and Newsql Databases Noname manuscript No. (will be inserted by the editor) The SQL++ Unifying Semi-structured Query Language, and an Expressiveness Benchmark of SQL-on-Hadoop, NoSQL and NewSQL Databases Kian Win Ong · Yannis Papakonstantinou · Romain Vernoux Received: date / Accepted: date Abstract SQL-on-Hadoop, NewSQL and NoSQL Since SQL itself is a subset of SQL++, the SQL-aware databases provide semi-structured data models (typ- reader will easily identify in which ways each of the ically JSON based) and respective query languages. surveyed databases provides more or less than SQL. Lack of formal syntax and semantics, idiomatic (non- The eleven databases are Hive, Jaql, Pig, Cassandra, SQL) language constructs and large variations in syn- JSONiq, MongoDB, Couchbase, SQL, AsterixDB, Big- tax, semantics and actual capabilities pose problems Query and UnityJDBC. They were selected due to their even to database experts: It is hard to understand, com- market adoption or because they present cutting edge, pare and use these languages. It is especially tedious to advanced query language abilities. write software that interoperates between two of them Finally, we briefly discuss the use of SQL++ as or an SQL database and one of them. the query language of the FORWARD virtual database Towards solving these problems, first we formally query processor, which executes SQL++ queries over specify the syntax and semantics of SQL++. It con- SQL and non-SQL databases and the use of SQL++ in sists of a semi-structured data model (which extends the FORWARD application framework, which enables both JSON and the relational data model) and a query rapid development of live reports and interactive appli- language that is fully backwards compatible with SQL. cations on SQL and non-SQL databases. FORWARD SQL++ is “unifying” in the sense that it is explicitly provides a proof-of-concept of SQL++’s applicability designed to encompass the data model and query lan- as a unifying data model and query language. guage capabilities of current SQL-on-Hadoop, NoSQL and NewSQL databases. Then, we itemize fifteen SQL++ data model and query language features and benchmark eleven 1 Introduction databases on their support of the multiple options as- sociated with each feature, leading to feature matri- Numerous databases marketed as SQL-on-Hadoop, ces and commentary. Each feature matrix is the re- NewSQL [30] and NoSQL have emerged to catalyze Big sult of empirical validation through sample queries. Data applications. These databases generally support the 3Vs [11]. (i) Volume: amount of data (ii) Velocity: This work was supported by NSF DC 0910820, NSF III 1219263, NSF speed of data in and out (iii) Variety: semi-structured IIS 1237174 and Informatica grants. The grants' PI is Prof. and heterogeneous data. As a result of differing use Papakonstantinou who is a shareholder of an entity that commercializes cases and design considerations around the Variety re- some results mentioned in this research. quirement, these new databases have adopted semi- Kian Win Ong structured data models that vary among each other. E-mail: [email protected] Their query languages have even more variations. Some Yannis Papakonstantinou∗ variations are due to superficial syntactic differences. E-mail: [email protected] Some variations arise from the data model differences. Romain Vernoux Finally, other variations are genuine differences in query E-mail: [email protected] capabilities. 2 Kian Win Ong et al. In this setting, even researchers and practitioners that push the agenda on query languages for JSON or with many years of SQL database experience face prob- JSON-like data, as these databases gravitate towards lems in two areas: more sophisticated and complete query capabilities. The benchmark’s results are presented through 1. Comprehension: Significant effort is needed to un- fifteen feature matrices and additional analy- derstand, compare and contrast the semi-structured sis/commentary that classify each database’s data data models and query languages of these novel model and query language capabilities as a subset of databases. The informal (and often underspecified) SQL++. The matrices further decompose each feature syntax and semantics of the provided query languages into as many as eleven constituent sub-features and make comprehension even harder or impossible, as it options, in order to facilitate fine-grained comparisons becomes apparent from the avalanche of syntax and across different data models and languages. Besides semantics questions in online forums. providing information on supported and unsupported 2. Development: It is difficult to write software that re- features and options, the matrices also qualify capa- trieves data from multiple such databases, given the bility differences that cut across individual features, different data models, different query syntaxes and such as the composability of various query language the (often subtly) different query semantics. These features with each other. For readability, we interleave interoperability issues occur frequently in practice, the SQL++ specification sections with the respective for example, whenever an organization adopts one benchmarking (capability classification) sections. of these new databases and then builds applications The approach of outlining the differences between that need integrated access to data stored in the new the various databases using SQL++ achieves two ben- database and in its existing SQL databases. efits: First, SQL++ offers the reader a formal specifica- tion of the discussed features and capabilities. Second, Towards solving the above problems, we formally by understanding each database’s capabilities in terms specify the syntax and semantics of SQL++, which is of SQL++, the reader can focus on the fundamental a unifying semi-structured data model and query lan- differences of the databases without being confused by guage that is designed to encompass the data model syntactic idiosyncracies of various query languages and and query language capabilities of NoSQL, NewSQL superficial differences in the documented descriptions and SQL-on-Hadoop databases. The SQL++ semantics of their semantics. stands on the shoulders of the extensive past work from The relatively immature state of query language the database R&D community in non-relational data documentation of the surveyed databases leaves many models and query languages: OQL [2], the nested rela- questions unanswered. We dealt with this problem us- tional model and query languages [15,28,1] and XQuery ing a hands-on approach: Each feature matrix has been (and other XML-based query languages) [27,10,5]. empirically validated by executing sample queries on SQL++ is an extension to SQL and is backwards- the surveyed databases. A benchmark comprising sam- compatible with SQL. This choice was made in order ple queries, empirical observations, as well as links to to facilitate the SQL-aware audience in two aspects: supporting documentation and bug reports is available First, since many surveyed databases do not support at . the entirety of standard SQL capabilities, the provided http://forward.ucsd.edu/sqlpp The feature matrices of this survey paper classify comparisons explain the extent to which each surveyed many capabilities of semi-structured data models and database supports the SQL capabilities. Second and query languages. The most prominent capabilities are: most importantly, the reader will understand in what ways semi-structured data models and query languages – What kinds of data values are supported by each extend SQL’s capabilities, understand in which ways database? these extensions may relate to each other, and obtain – What kind of schemas and constraints are supported? an overview on which surveyed databases support these extensions. – How does the query language access and construct Then we itemize fifteen SQL++ data model nested data? and query language features and benchmark eleven – How is missing information represented and handled? databases on their support of the multiple options as- – What are the options and semantics for equality on sociated with each feature. For this benchmark, we non-scalar and heterogeneous values? cover the most popular SQL-on-Hadoop, NoSQL and NewSQL databases from DB-Engines [8] (a popularity – What are the options and semantics for ordering on tracker for database engines) and industry surveys [12, non-scalar and heterogeneous values? 17]. We have also selected research-oriented databases – Is aggregation supported? The SQL++ Semi-structured Query Language 3 – Is join supported? <=;#4/& – Are extensions (such as UDFs) provided to circum- !"#$$% !"#$$% vent limitations? :"#$;#*& $#*"=/*& &'()*(+%,-../01230& We expect that some of the results listed in the feature matrices will change in the next years as the !"#$$%!"#$%&'$()#**($& surveyed databases will release newer and better im- plementations. The arXiv/CoRR version of this paper !"#$$&@;$/".=&-./.0.*#& [24] and the benchmark will be updated to reflect these changes. +!,??& +!,??& +!,??& +!,??& +!,??& @;$/".=& @;$/".=& @;$/".=& @;$/".=& @;$/".=& Despite the forthcoming changes, we expect SQL++ @;#2*& @;#2*& @;#2*& @;#2*& @;#2*& and the comparison methodology followed by this sur- +!,3(43 +!,& 1#2+!,& 1(+!,& 5.6((7& >>>& vey to remain a standing contribution. Besides its value A$.77#$& A$.77#$& A$.77#$& A$.77#$& to developers, SQL++ can also assist the query lan- guage designers in the NoSQL, NewSQL and SQL-on- 1.89#& 1.89#& Hadoop space towards (1) producing formal versions
Recommended publications
  • Towards an Analytics Query Engine *
    Towards an Analytics Query Engine ∗ Nantia Makrynioti Vasilis Vassalos Athens University of Economics and Business Athens University of Economics and Business Athens, Greece Athens, Greece [email protected] [email protected] ABSTRACT with black box libraries, evaluating various algorithms for a This vision paper presents new challenges and opportuni- task and tuning their parameters, in order to produce an ef- ties in the area of distributed data analytics, at the core of fective model, is a time-consuming process. Things become which are data mining and machine learning. At first, we even more complicated when we want to leverage paralleliza- provide an overview of the current state of the art in the area tion on clusters of independent computers for processing big and then analyse two aspects of data analytics systems, se- data. Details concerning load balancing, scheduling or fault mantics and optimization. We argue that these aspects will tolerance can be quite overwhelming even for an experienced emerge as important issues for the data management com- software engineer. munity in the next years and propose promising research Research in the data management domain recently started directions for solving them. tackling the above issues by developing systems for large- scale analytics that aim at providing higher-level primitives for building data mining and machine learning algorithms, Keywords as well as hiding low-level details of distributed execution. Data analytics, Declarative machine learning, Distributed MapReduce [12] and Dryad [16] were the first frameworks processing that paved the way. However, these initial efforts suffered from low usability, as they offered expressive but at the same 1.
    [Show full text]
  • Evaluation of Xpath Queries on XML Streams with Networks of Early Nested Word Automata Tom Sebastian
    Evaluation of XPath Queries on XML Streams with Networks of Early Nested Word Automata Tom Sebastian To cite this version: Tom Sebastian. Evaluation of XPath Queries on XML Streams with Networks of Early Nested Word Automata. Databases [cs.DB]. Universite Lille 1, 2016. English. tel-01342511 HAL Id: tel-01342511 https://hal.inria.fr/tel-01342511 Submitted on 6 Jul 2016 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Universit´eLille 1 – Sciences et Technologies Innovimax Sarl Institut National de Recherche en Informatique et en Automatique Centre de Recherche en Informatique, Signal et Automatique de Lille These` pr´esent´ee en premi`ere version en vue d’obtenir le grade de Docteur par l’Universit´ede Lille 1 en sp´ecialit´eInformatique par Tom Sebastian Evaluation of XPath Queries on XML Streams with Networks of Early Nested Word Automata Th`ese soutenue le 17/06/2016 devant le jury compos´ede : Carlo Zaniolo University of California Rapporteur Anca Muscholl Universit´eBordeaux Rapporteur Kim Nguyen Universit´eParis-Sud Examinateur Remi Gilleron Universit´eLille 3 Pr´esident du Jury Joachim Niehren INRIA Directeur Abstract The eXtended Markup Language (Xml) provides a format for representing data trees, which is standardized by the W3C and largely used today for exchanging information between all kinds of computer programs.
    [Show full text]
  • QUERYING JSON and XML Performance Evaluation of Querying Tools for Offline-Enabled Web Applications
    QUERYING JSON AND XML Performance evaluation of querying tools for offline-enabled web applications Master Degree Project in Informatics One year Level 30 ECTS Spring term 2012 Adrian Hellström Supervisor: Henrik Gustavsson Examiner: Birgitta Lindström Querying JSON and XML Submitted by Adrian Hellström to the University of Skövde as a final year project towards the degree of M.Sc. in the School of Humanities and Informatics. The project has been supervised by Henrik Gustavsson. 2012-06-03 I hereby certify that all material in this final year project which is not my own work has been identified and that no work is included for which a degree has already been conferred on me. Signature: ___________________________________________ Abstract This article explores the viability of third-party JSON tools as an alternative to XML when an application requires querying and filtering of data, as well as how the application deviates between browsers. We examine and describe the querying alternatives as well as the technologies we worked with and used in the application. The application is built using HTML 5 features such as local storage and canvas, and is benchmarked in Internet Explorer, Chrome and Firefox. The application built is an animated infographical display that uses querying functions in JSON and XML to filter values from a dataset and then display them through the HTML5 canvas technology. The results were in favor of JSON and suggested that using third-party tools did not impact performance compared to native XML functions. In addition, the usage of JSON enabled easier development and cross-browser compatibility. Further research is proposed to examine document-based data filtering as well as investigating why performance deviated between toolsets.
    [Show full text]
  • Programming a Parallel Future
    Programming a Parallel Future Joseph M. Hellerstein Electrical Engineering and Computer Sciences University of California at Berkeley Technical Report No. UCB/EECS-2008-144 http://www.eecs.berkeley.edu/Pubs/TechRpts/2008/EECS-2008-144.html November 7, 2008 Copyright 2008, by the author(s). All rights reserved. Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, to republish, to post on servers or to redistribute to lists, requires prior specific permission. Programming a Parallel Future Joe Hellerstein UC Berkeley Computer Science Things change fast in computer science, but odds are good that they will change especially fast in the next few years. Much of this change centers on the shift toward parallel computing. In the short term, parallelism will thrive in settings with massive datasets and analytics. Longer term, the shift to parallelism will impact all software. In this note, I’ll outline some key changes that have recently taken place in the computer industry, show why existing software systems are ill‐equipped to handle this new reality, and point toward some bright spots on the horizon. Technology Trends: A Divergence in Moore’s Law Like many changes in computer science, the rapid drive toward parallel computing is a function of technology trends in hardware. Most technology watchers are familiar with Moore's Law, and the general notion that computing performance per dollar has grown at an exponential rate — doubling about every 18‐24 months — over the last 50 years.
    [Show full text]
  • Declarative Data Analytics: a Survey
    Declarative Data Analytics: a Survey NANTIA MAKRYNIOTI, Athens University of Economics and Business, Greece VASILIS VASSALOS, Athens University of Economics and Business, Greece The area of declarative data analytics explores the application of the declarative paradigm on data science and machine learning. It proposes declarative languages for expressing data analysis tasks and develops systems which optimize programs written in those languages. The execution engine can be either centralized or distributed, as the declarative paradigm advocates independence from particular physical implementations. The survey explores a wide range of declarative data analysis frameworks by examining both the programming model and the optimization techniques used, in order to provide conclusions on the current state of the art in the area and identify open challenges. CCS Concepts: • General and reference → Surveys and overviews; • Information systems → Data analytics; Data mining; • Computing methodologies → Machine learning approaches; • Software and its engineering → Domain specific languages; Data flow languages; Additional Key Words and Phrases: Declarative Programming, Data Science, Machine Learning, Large-scale Analytics ACM Reference Format: Nantia Makrynioti and Vasilis Vassalos. 2019. Declarative Data Analytics: a Survey. 1, 1 (February 2019), 36 pages. https://doi.org/10.1145/nnnnnnn.nnnnnnn 1 INTRODUCTION With the rapid growth of world wide web (WWW) and the development of social networks, the available amount of data has exploded. This availability has encouraged many companies and organizations in recent years to collect and analyse data, in order to extract information and gain valuable knowledge. At the same time hardware cost has decreased, so storage and processing of big data is not prohibitively expensive even for smaller companies.
    [Show full text]
  • A Platform-Independent Framework for Managing Geo-Referenced JSON Data Sets
    electronics Article J-CO: A Platform-Independent Framework for Managing Geo-Referenced JSON Data Sets Giuseppe Psaila * and Paolo Fosci Department of Management, Information and Production Engineering, University of Bergamo, 24044 Dalmine (BG), Italy; [email protected] * Correspondence: [email protected]; Tel.: +39-035-205-2355 Abstract: Internet technology and mobile technology have enabled producing and diffusing massive data sets concerning almost every aspect of day-by-day life. Remarkable examples are social media and apps for volunteered information production, as well as Open Data portals on which public administrations publish authoritative and (often) geo-referenced data sets. In this context, JSON has become the most popular standard for representing and exchanging possibly geo-referenced data sets over the Internet.Analysts, wishing to manage, integrate and cross-analyze such data sets, need a framework that allows them to access possibly remote storage systems for JSON data sets, to retrieve and query data sets by means of a unique query language (independent of the specific storage technology), by exploiting possibly-remote computational resources (such as cloud servers), comfortably working on their PC in their office, more or less unaware of real location of resources. In this paper, we present the current state of the J-CO Framework, a platform-independent and analyst-oriented software framework to manipulate and cross-analyze possibly geo-tagged JSON data sets. The paper presents the general approach behind the Framework, by illustrating J-CO the query language by means of a simple, yet non-trivial, example of geographical cross-analysis. The paper also presents the novel features introduced by the re-engineered version of the execution Citation: Psaila, G.; Fosci, P.
    [Show full text]
  • XML Prague 2012
    XML Prague 2012 Conference Proceedings University of Economics, Prague Prague, Czech Republic February 10–12, 2012 XML Prague 2012 – Conference Proceedings Copyright © 2012 Jiří Kosek ISBN 978-80-260-1572-7 Table of Contents General Information ..................................................................................................... vii Sponsors ........................................................................................................................... ix Preface .............................................................................................................................. xi The eX Markup Language? – Eric van der Vlist ........................................................... 1 XML and HTML Cross-Pollination: A Bridge Too Far? – Norman Walsh and Robin Berjon .................................................................................... 11 XML5's Story – Anne van Kesteren ................................................................................ 23 XProc: Beyond application/xml – Vojtěch Toman ........................................................ 27 The Anatomy of an Open Source XProc/XSLT implementation of NVDL – George Bina ...................................................................................................................... 49 JSONiq – Jonathan Robie, Matthias Brantner, Daniela Florescu, Ghislain Fourny, and Till Westmann .................................................................................................................. 63 Corona: Managing and
    [Show full text]
  • Big Data Analytic Approaches Classification
    Big Data Analytic Approaches Classification Yudith Cardinale1 and Sonia Guehis2,3 and Marta Rukoz2,3 1Dept. de Computacion,´ Universidad Simon´ Bol´ıvar, Venezuela 2Universite´ Paris Nanterre, 92001 Nanterre, France 3Universite´ Paris Dauphine, PSL Research University, CNRS, UMR[7243], LAMSADE, 75016 Paris, France Keywords: Big Data Analytic, Analytic Models for Big Data, Analytical Data Management Applications. Abstract: Analytical data management applications, affected by the explosion of the amount of generated data in the context of Big Data, are shifting away their analytical databases towards a vast landscape of architectural solutions combining storage techniques, programming models, languages, and tools. To support users in the hard task of deciding which Big Data solution is the most appropriate according to their specific requirements, we propose a generic architecture to classify analytical approaches. We also establish a classification of the existing query languages, based on the facilities provided to access the Big Data architectures. Moreover, to evaluate different solutions, we propose a set of criteria of comparison, such as OLAP support, scalability, and fault tolerance support. We classify different existing Big Data analytics solutions according to our proposed generic architecture and qualitatively evaluate them in terms of the criteria of comparison. We illustrate how our proposed generic architecture can be used to decide which Big Data analytic approach is suitable in the context of several use cases. 1 INTRODUCTION nally, despite these increases in scale and complexity, users still expect to be able to query data at interac- The term Big Data has been coined for represent- tive speeds. In this context, several enterprises such ing the challenge to support a continuous increase as Internet companies and Social Network associa- on the computational power that produces an over- tions have proposed their own analytical approaches, whelming flow of data (Kune et al., 2016).
    [Show full text]
  • Download the Result from the Cluster
    UC Riverside UC Riverside Electronic Theses and Dissertations Title Interval Joins for Big Data Permalink https://escholarship.org/uc/item/7xb001cz Author Carman, Jr., Eldon Preston Publication Date 2020 License https://creativecommons.org/licenses/by/4.0/ 4.0 Peer reviewed|Thesis/dissertation eScholarship.org Powered by the California Digital Library University of California UNIVERSITY OF CALIFORNIA RIVERSIDE Interval Joins for Big Data A Dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in Computer Science by Eldon Preston Carman, Jr. September 2020 Dissertation Committee: Dr. Vassilis J. Tsotras, Chairperson Dr. Michael J. Carey Dr. Ahmed Eldawy Dr. Vagelis Hristidis Dr. Eamonn Keogh Copyright by Eldon Preston Carman, Jr. 2020 The Dissertation of Eldon Preston Carman, Jr. is approved: Committee Chairperson University of California, Riverside Acknowledgments I am grateful to my advisor, Professor Vassilis Tsotras, without whose support, I would not have been able to complete this journey. Thank you for your continued encour- agement and guidance through this learning experience. I especially would like to thank my dissertation committee members, Dr. Ahmed Eldawy, Dr. Vagelis Hristidis, Dr. Eamonn Keogh, and Dr. Michael Carey, for reviewing my dissertation. I also like to thank Dr. Rajiv Gupta for his support in the beginning of this dissertation. I would like to thank everyone in AsterixDB’s team for their support. Especially Professor Michael Carey for his guidance in understanding the AsterixDB stack and managing the research in an active software product. Thank you to my lab partners, Ildar Absalyamov and Steven Jacobs, for all the lively discussions on our regular drives out to Irvine.
    [Show full text]
  • A Dataflow Language for Large Scale Processing of RDF Data
    SYRql: A Dataflow Language for Large Scale Processing of RDF Data Fadi Maali1, Padmashree Ravindra2, Kemafor Anyanwu2, and Stefan Decker1 1 Insight Centre for Data Analytics, National University of Ireland Galway {fadi.maali,stefan.decker}@insight-centre.org 2 Department of Computer Science, North Carolina State University, Raleigh, NC {pravind2,kogan}@ncsu.edu Abstract. The recent big data movement resulted in a surge of activity on layering declarative languages on top of distributed computation plat- forms. In the Semantic Web realm, this surge of analytics languages was not reflected despite the significant growth in the available RDF data. Consequently, when analysing large RDF datasets, users are left with two main options: using SPARQL or using an existing non-RDF-specific big data language, both with its own limitations. The pure declarative nature of SPARQL and the high cost of evaluation can be limiting in some scenarios. On the other hand, existing big data languages are de- signed mainly for tabular data and, therefore, applying them to RDF data results in verbose, unreadable, and sometimes inefficient scripts. In this paper, we introduce SYRql, a dataflow language designed to process RDF data at a large scale. SYRql blends concepts from both SPARQL and existing big data languages. We formally define a closed algebra that underlies SYRql and discuss its properties and some unique optimisation opportunities this algebra provides. Furthermore, we describe an imple- mentation that translates SYRql scripts into a series of MapReduce jobs and compare the performance to other big data processing languages. 1 Introduction Declarative query languages have been a corner stone of data management since the early days of relational databases.
    [Show full text]
  • Recent Trends in JSON Filters Atul Jain*, Dr
    International Journal of Scientific Research in Computer Science, Engineering and Information Technology ISSN : 2456-3307 (www.ijsrcseit.com) https://doi.org/10.32628/CSEIT217116 Recent Trends in JSON Filters Atul Jain*, Dr. ShashiKant Gupta CSA Department, ITM University, Gwalior, Madhya Pradesh, India ABSTRACT Article Info JavaScript Object Notation is a text-based data exchange format for structuring Volume 7, Issue 1 data between a server and web application on the client-side. It is basically a Page Number: 87-93 data format, so it is not limited to Ajax-style web applications and can be used Publication Issue : with API’s to exchange or store information. However, the whole data never to January-February-2021 be used by the system or application, It needs some extract of a piece of requirement that may vary person to person and with the changing of time. The searching and filtration from the JSON string are very typical so most of the studies give only basics operation to query the data from the JSON object. The aim of this paper to find out all the methods with different technology to search and filter with JSON data. It explains the extensive results of previous research on the JSONiq Flwor expression and compares it with the json-query module of npm to extract information from JSON. This research has the intention of achieving the data from JSON with some advanced operators with the help of a prototype in json-query package of NodeJS. Thus, the data can be filtered out more efficiently and accurately Article History without the need for any other programming language dependency.
    [Show full text]
  • Extracting Data from Nosql Databases a Step Towards Interactive Visual Analysis of Nosql Data Master of Science Thesis
    Extracting Data from NoSQL Databases A Step towards Interactive Visual Analysis of NoSQL Data Master of Science Thesis PETTER NÄSHOLM Chalmers University of Technology University of Gothenburg Department of Computer Science and Engineering Göteborg, Sweden, January 2012 The Author grants to Chalmers University of Technology and University of Gothenburg the non-exclusive right to publish the Work electronically and in a non-commercial purpose make it accessible on the Internet. The Author warrants that he/she is the author to the Work, and warrants that the Work does not contain text, pictures or other material that violates copyright law. The Author shall, when transferring the rights of the Work to a third party (for example a publisher or a company), acknowledge the third party about this agreement. If the Author has signed a copyright agreement with a third party regarding the Work, the Author warrants hereby that he/she has obtained any necessary permission from this third party to let Chalmers University of Technology and University of Gothenburg store the Work electronically and make it accessible on the Internet. Extracting Data from NoSQL Databases A Step towards Interactive Visual Analysis of NoSQL Data PETTER NÄSHOLM © PETTER NÄSHOLM, January 2012. Examiner: GRAHAM KEMP Chalmers University of Technology University of Gothenburg Department of Computer Science and Engineering SE-412 96 Göteborg Sweden Telephone + 46 (0)31-772 1000 Department of Computer Science and Engineering Göteborg, Sweden January 2012 Abstract Businesses and organizations today generate increasing volumes of data. Being able to analyze and visualize this data to find trends that can be used as input when making business decisions is an important factor for competitive advan- tage.
    [Show full text]