The Potential of Temporal Databases for the Application in Data Analytics

Total Page:16

File Type:pdf, Size:1020Kb

The Potential of Temporal Databases for the Application in Data Analytics The Potential of Temporal Databases for the Application in Data Analytics Master Thesis Alexander Menne July 2019 Thesis supervisors: Prof. Dr. Marko van Eekelen ICIS, Radboud University and Ronald van Herwijnen Avanade Netherlands Second assessor: Dr. Stijn Hoppenbrouwers ICIS, Radboud University Radboud University and Avanade Netherlands Abstract The concept of temporal databases has gotten much attention from research in computer science, but there is a lack of literature concerning the practical application of temporal databases. This thesis examines the potential of temporal databases for data analyt- ics. The main concepts of temporal databases and the role of temporal data for data analytics and data warehousing is studied using a literature review. Furthermore, the current implementations of temporal databases are discussed to exemplify the differences between the literature and practice. We compare temporal databases embedded in a data warehouse architecture with conventional data warehouses by means of two proto- types and five assessment criteria. The results of the assessment indicate that the use of temporal databases as alternative for traditional data warehouses has great advantages. Temporal databases solve data integrity issues of the classical ETL process and enable a more direct data flow from the source database to the business intelligence tool. Also, the integrated support of the temporal dimension reduces programming efforts and in- creases the maintainability of the system. Hence, we find that temporal databases have the potential to enhance the data-driven strategy of companies significantly. Keywords: Temporal Database, Data Analytics, Data Warehouse, ETL 1 Acknowledgements I would like to thank Marko van Eekelen for the excellent guidance through our biweekly meetings which provided me with very valuable insights into research methodology. Also, I want to thank my external supervisor Ronald van Herwijnen for sharing his expertise in data warehousing, which contributed significantly to the prototype construction. Fur- thermore, I want to thank Stijn Hoppenbrouwers for the time and effort he has devoted to reading and assessing this thesis. Finally, I want to express my gratitude to all the people that supported, encouraged, and inspired me on this academic journey in the Netherlands, which is a place I am proud to call home. 3 Table of Contents 1. Introduction7 1.1. Related work . .8 1.2. Research question . .9 1.3. Thesis outline . 10 2. Research Methods 11 2.1. Literature review . 11 2.2. Assessment . 12 2.2.1. Issues and criteria . 12 2.2.2. Measurement . 13 2.2.3. Technology . 15 3. Theoretical Background 19 3.1. Temporal Data . 19 3.2. Conventional Databases and Temporal Data . 20 3.3. Temporal Databases . 22 3.3.1. Implementation issues . 24 3.4. Data Analytics . 25 3.4.1. Big Data . 26 3.4.2. The role of temporal data . 27 3.5. Data Warehouses . 27 3.5.1. The temporal dimension . 29 4. Practical Background 31 4.1. SQL:2011 standard . 31 4.2. Microsoft SQL Server implementation . 35 4.3. Other implementations . 37 5 6 Table of Contents 5. Temporal Databases compared to Conventional Data Warehouses 39 5.1. Prototype A: Conventional data warehouse . 40 5.1.1. Architecture . 40 5.1.2. Insights . 40 5.2. Prototype B: Data warehouse with system-versioned tables . 41 5.2.1. Architecture . 41 5.2.2. Insights . 42 5.3. Assessment results . 42 5.3.1. Performance . 43 5.3.2. Costs . 44 5.3.3. Data integrity . 45 5.3.4. Maintainability . 46 5.3.5. Acceptance . 47 5.3.6. Three issues of the classical ETL process . 47 6. Discussion 49 6.1. Interpretation of the findings . 49 6.2. Limitations . 52 7. Conclusion 53 7.1. Conclusions . 53 7.2. Future work . 54 Glossary 55 References 57 A. Source Code 61 A.1. Transfer between staging database and data warehouse . 61 A.2. Data warehouse transformation . 70 A.3. Views in temporal database . 98 B. Technical specifications 115 C. Assessment 117 C.1. Performance assessment results . 117 C.2. Assessment scripts . 119 1 Introduction 1.1. Related work ........................... 8 1.2. Research question ......................... 9 1.3. Thesis outline ........................... 10 Thirty years ago, the British computer scientist Tim Berners-Lee submitted a ’vague but exciting’ proposal for an information system meant as a "free, open, permission- less space for all of humanity to share knowledge and ideas" [32]. Nowadays, we call this system the internet, and its purpose has expanded far beyond the spreading of knowledge. All kinds of data are stored online, which leads to the exponential growth of the internet. The fact that the data stored online doubles every 20 months gives an indication of the challenge that companies face to cope with the masses of collected data [26]. The field of data analytics aims at finding solutions for the increasingly important extraction of knowledge to empower data-driven company strategies. The internet has become not only a source of data but also a facilitator for data analytics due to the efficiency and flexibility that cloud computing offers. In spite of the technological progress driven by the internet, the architecture of data warehouses has hardly changed in the last two decades. Databases usually represent the current state of an organisation and are updated when a change happens in the real world. Periodically, data is extracted, transformed, and loaded (ETL) into a data warehouse to gain insights for decision-making processes. In between two ETL processes, the propositions stored in the database can change multiple times, which is not reflected in the next ETL process, as only the current state of the database is extracted at the start of the ETL process. In this fast-changing world, this is a potential bias for the decision-making process, since ETL processes are often executed only once per day or less. Temporal databases may solve this problem by recording all changes made to the database in history tables. Hence, no data ever gets lost as the archiving of the data is not dependent from an ETL process. This, however, also means that much data needs to be stored, which might be a reason why it took almost 30 years from the first idea 7 8 Chapter 1. Introduction Fig. 1.1.: Average disk drive price per gigabyte in US dollar, based on [21]. of a temporal database to an implementation by a database management system ven- dor. Nowadays, the conditions have changed as the prices for hardware have declined immensely, which makes it affordable to store large amounts of data even for small busi- nesses. An illustration of the price decline of disk drives from 2004 to 2019 can be found in figure 1.1. Surprisingly, the low storage prices and available solutions has not led to an establishment of temporal databases in the industry yet. 1.1. Related work Much research on temporal databases has been done in the field of computer science, which mainly focuses on the underlying concepts and possible implementations of tem- poral databases in existing database management systems. With over 100 publications, Richard T. Snodgrass is the primary contributor to the design and implementation of temporal databases [31]. His early publications on temporal databases contribute for a significant part to the theoretical foundation this thesis is built on [10, 9]. Also, Snod- grass is co-director of the institution TimeCenter, which plays an essential role in the advancement of knowledge within the domain of temporal databases [17]. More recent works include the books of Johnston [18] and Date et al. [14], which combine the theo- retical aspects of temporal data with a rather practical view on the design of temporal databases. Especially, Date et al. [14] offer an interesting discussion on the problems that arise when implementing a temporal database. There is a relatively small body of literature that is solely concerned with data analytics in general. Most recent academic research focuses on trends within the field of data analytics, such as big data. Gandomi et al. [2] provide a comprehensive overview of relevant techniques in the field of big data analytics. The study of Russom [27] offers some important insights into the state of big data analytics and best practices of the 1.2. Research question 9 industry as reported by 325 data management professionals. However, the reader should bear in mind that the study was taken in 2011. Closely related to data analytics, data warehouses are foremost addressed by practical researchers and industry leaders. Ralph Kimball and the Kimball Group are important authorities in this field, as they established the best practices of data warehouse archi- tecture with their ’Toolkit’ books [20]. A main contribution of Kimball is the concept of dimensional modelling, which implies that tables in the data warehouse are modelled as a star schema with fact tables surrounded by dimension tables [19]. In contrast to the data warehouse design suggested by Inmon [16], the Kimball data warehouse is not normalised to simplify the information retrieval process. Overall, there is sufficient theoretical research on temporal databases and practical re- search on data analytics and data warehousing. There is, however, very little known about the use of temporal databases for the purposes of data analytics. Also, the use of a temporal database within a data warehouse architecture has not been investigated yet. 1.2. Research question This research aims at closing the knowledge gap on the use of temporal databases in data warehouse architectures to achieve
Recommended publications
  • LINGI2172 Databases 2013-2014
    Université Catholique de Louvain - DESCRIPTIF DE COURS 2013-2014 - LINGI2172 LINGI2172 Databases 2013-2014 6.0 crédits 30.0 h + 30.0 h 2q Enseignants: Lambeau Bernard ; Langue Français d'enseignement: Lieu du cours Louvain-la-Neuve Ressources en ligne: > http://icampus.uclouvain.be/claroline/course/index.php?cid=lingi2172 Préalables : Basic knowledge of database management, good abilities in programming. Thèmes abordés : * Data Base Management Systems (objectives, requirements, architecture). * The Relational data model (formal theory, first-order logic, constraints). * Conceptual models (entity-relationship, object role modeling). * Logical database design (normal forms & mp; normalization, ER-To-Relational) * Physical database design and storage (tables and keys, indexes, file structures). * Querying databases (Relational Algebra, Relational Calculus, Tutorial D, SQL) * ACID properties (Atomicity, Consistency, Isolation, Durability), Concurrency Control, Recovery techniques. * Programming database applications (JDBC, Database Cursors, Object-Relational Mapping, Relations as First-class Citizen). * Recent or more advanced trends in the database field (object-oriented databases, Big Data, NoSQL, NewSQL) Acquis Students completing this course successfully will be able to : * Explain the scenarios in which using a database is more convenient than programming with data files; d'apprentissage * Explain the characteristics of the database approach, where they come from and contrast them with current trends in the database field-- Identify and describe the main functions of a database management system; * Categorize conceptual, logical and physical data models based on the concepts they provide to describe the database structure; * Understand the main principles and mathematical theory of the relational approach to database management; * Design databases using a systematic approach, from a conceptual model through a logical level (i.e., a relational schema) into a physical model (i.e., tables and indexes); * Use SQL (DDL) to implement a relational database schema.
    [Show full text]
  • A Logical Framework for Temporal Deductive Databases S
    A logical framework for temporal deductive databases S. M. Sripada Departmentof Computing Imperial College of Science& Technology 180 Queen’sGate, London SW7 2BZ Abstract researchers have incorporated time into conventional databasesusing various schemes Temporal deductive databases are deductive providing different capabilities for handling databaseswith an ability to representboth valid temporal information. In particular, work has time and transaction time. The work is basedon been done on adding temporal information to the Event Calculus of Kowalski & Sergot. Event conventionaldatabases to turn them into temporal Calculus is a treatment of time, based on the databases. However, no work is done on notion of events, in first-order classical logic handling both transaction time and valid time in augmentedwith negation as failure. It formalizes deductive databases.In this paper we propose a the semanticsof valid time in deductivedatabases framework for dealing with time in temporal and offers capability for the semantic validation deductivedatabases. of updates and default reasoning. In this paper, the Event Calculus is extended to include the Snodgrass dz Ahn [16,17] describe a concept of transaction time. The resulting classification of databasesdepending on their framework is capable of handling both proactive ability to represent temporal information. They and retroactive updates symmetrically. Error identify three different concepts of time - valid correction is achievedwithout deletionsby means time, transaction time and user-defined time. Of of negation as failure. The semantics of these, user-defined time is temporal transaction time is formalised and the axioms of information of some kind that is of interest to the the Event Calculus are modified to cater for user but does not concern a DBMS.
    [Show full text]
  • Table of Contents
    The Kimball Group Reader Relentlessly Practical Tools for Data Warehousing and Business Intelligence Remastered Collection Ralph Kimball and Margy Ross with Bob Becker, Joy Mundy, and Warren Thornthwaite Contents Introduction . xxv 1 The Reader at a Glance . 1 Setting Up for Success . 1 1.1 Resist the Urge to Start Coding . 1 1.2 Set Your Boundaries . 4 Tackling DW/BI Design and Development . 6 1.3 Data Wrangling . 6 1.4 Myth Busters . 9 1.5 Dividing the World . 10 1.6 Essential Steps for the Integrated Enterprise Data Warehouse . 13 1.7 Drill Down to Ask Why . 22 1.8 Slowly Changing Dimensions . 25 1.9 Judge Your BI Tool through Your Dimensions . 28 1.10 Fact Tables . 31 1.11 Exploit Your Fact Tables . 33 2 Before You Dive In . 35 Before Data Warehousing . 35 2.1 History Lesson on Ralph Kimball and Xerox PARC. 36 Historical Perspective . 37 2.2 The Database Market Splits . 37 2.3 Bringing Up Supermarts . 40 Dealing with Demanding Realities . 47 2.4 Brave New Requirements for Data Warehousing . 47 2.5 Coping with the Brave New Requirements. 52 2.6 Stirring Things Up . 57 2.7 Design Constraints and Unavoidable Realities . 60 xiv Contents 2.8 Two Powerful Ideas . 64 2.9 Data Warehouse Dining Experience . 67 2.10 Easier Approaches for Harder Problems . 70 2.11 Expanding Boundaries of the Data Warehouse . 72 3 Project/Program Planning . 75 Professional Responsibilities . 75 3.1 Professional Boundaries . 75 3.2 An Engineer’s View . 78 3.3 Beware the Objection Removers .
    [Show full text]
  • Temporal Metadata for Discovery
    Temporal Metadata for Discovery A review of current practice Makx Dekkers AMI Consult SARL Massimo Craglia (Editor) European Commission DG JRC Institute for Environment and Sustainability EUR 23209 EN - 2008 The mission of the Institute for Environment and Sustainability is to provide scientific-technical support to the European Union’s Policies for the protection and sustainable development of the European and global environment. European Commission Joint Research Centre Institute for Environment and Sustainability Contact information Makx Dekkers Address: AMI Consult Société à responsabilité limitée B.P. 647 2016 Luxembourg E-mail: [email protected] Massimo Craglia (Editor) Address: European Commission Joint Research Centre Institute for Environment and Sustainability Spatial Data Infrastructures Unit TP262, Via Fermi 2749 I-21027 Ispra (VA) ITALY E-mail: [email protected] Tel.: +39-0332-786269 Fax: +39-0332-786325 http://ies.jrc.ec.europa.eu/ http://www.jrc.ec.europa.eu/ Legal Notice Neither the European Commission nor any person acting on behalf of the Commission is responsible for the use which might be made of this publication. Europe Direct is a service to help you find answers to your questions about the European Union Freephone number (*): 00 800 6 7 8 9 10 11 (*) Certain mobile telephone operators do not allow access to 00 800 numbers or these calls may be billed. A great deal of additional information on the European Union is available on the Internet. It can be accessed through the Europa server http://europa.eu/ JRC 42620 EUR 23209 EN ISSN 1018-5593 Luxembourg: Office for Official Publications of the European Communities © European Communities, 2008 Reproduction is authorised provided the source is acknowledged Printed in Italy Table of contents 1 BACKGROUND AND RATIONALE ....................................................................
    [Show full text]
  • Basically Speaking, Inmon Professes the Snowflake Schema While Kimball Relies on the Star Schema
    What is the main difference between Inmon and Kimball? Basically speaking, Inmon professes the Snowflake Schema while Kimball relies on the Star Schema. According to Ralf Kimball… Kimball views data warehousing as a constituency of data marts. Data marts are focused on delivering business objectives for departments in the organization. And the data warehouse is a conformed dimension of the data marts. Hence a unified view of the enterprise can be obtained from the dimension modeling on a local departmental level. He follows Bottom-up approach i.e. first creates individual Data Marts from the existing sources and then Create Data Warehouse. KIMBALL – First Data Marts – Combined way – Data warehouse. According to Bill Inmon… Inmon beliefs in creating a data warehouse on a subject-by-subject area basis. Hence the development of the data warehouse can start with data from their needs arise. Point-of-sale (POS) data can be added later if management decides it is necessary. He follows Top-down approach i.e. first creates Data Warehouse from the existing sources and then create individual Data Marts. INMON – First Data warehouse – Later – Data Marts. The Main difference is: Kimball: follows Dimensional Modeling. Inmon: follows ER Modeling bye Mayee. Kimball: creating data marts first then combining them up to form a data warehouse. Inmon: creating data warehouse then data marts. What is difference between Views and Materialized Views? Views: •• Stores the SQL statement in the database and let you use it as a table. Every time you access the view, the SQL statement executes. •• This is PSEUDO table that is not stored in the database and it is just a query.
    [Show full text]
  • MASTER's THESIS Role of Metadata in the Datawarehousing Environment
    2006:24 MASTER'S THESIS Role of Metadata in the Datawarehousing Environment Kranthi Kumar Parankusham Ravinder Reddy Madupu Luleå University of Technology Master Thesis, Continuation Courses Computer and Systems Science Department of Business Administration and Social Sciences Division of Information Systems Sciences 2006:24 - ISSN: 1653-0187 - ISRN: LTU-PB-EX--06/24--SE Preface This study is performed as the part of the master’s programme in computer and system sciences during 2005-06. It has been very useful and valuable experience and we have learned a lot during the study, not only about the topic at hand but also to manage to the work in the specified time. However, this workload would not have been manageable if we had not received help and support from a number of people who we would like to mention. First of all, we would like to thank our professor Svante Edzen for his help and supervision during the writing of thesis. And also we would like to express our gratitude to all the employees who allocated their valuable time to share their professional experience. On a personal level, Kranthi would like to thank all his family for their help and especially for my friends Kiran, Chenna Reddy, and Deepak Kumar. Ravi would like to give the greatest of thanks to his family for always being there when needed, and constantly taking so extremely good care of me….Also, thanks to all my friends for being close to me. Luleå University of Technology, 31 January 2006 Kranthi Kumar Parankusham Ravinder Reddy Madupu Abstract In order for a well functioning data warehouse to succeed many components must work together.
    [Show full text]
  • Kimball Vs. Inmon
    © 2011 - Andy Hogg Kimball vs. Inmon "Neither are any wars so furious and bloody, or of so long continuance as those occasioned by difference in opinion, especially if it be in things indifferent." (Swift, 1726) In the world of the data warehouse (DW) there are two dominant and opposing dogmas. Zealots of both extoll the virtues of their chosen doctrine with religious fervour, whilst decrying the beliefs of the other. These doctrines have existed for years, and in that time innumerable DWs have been built upon the principles of William Inmon. Likewise, an incalculable number built upon the ideas of Ralph Kimball. Inmon’s Corporate Information Factory (CIF), is a top-down approach. Since the whole DW is built in advance of usage, it requires significant time to deliver value. It therefore requires unwavering sponsorship from a senior figure within the organisation, possessing long-term vision of the DW’s value. Commentators contrast Kimball’s Bus Architecture (BA) as a bottom-up approach, where the data marts (DMs) are built first and unified into a DW at the end of the process. Inmon (n.d.a) ridicules this:- “…in bottom up data warehouse development first one data mart is developed, then another data mart is developed, then one day - presto - you magically and effortlessly wake up and have a data warehouse”. Kimball (2003) dislikes the bottom-up description, “Bottom-up is typically viewed as quick and dirty – focused on the needs of a single department rather than the enterprise”. He maintains the BA is a holistic view of the enterprise, with a final overall structure planned from the outset.
    [Show full text]
  • A Framework for Ontology-Based Library Data Generation, Access and Exploitation
    Universidad Politécnica de Madrid Departamento de Inteligencia Artificial DOCTORADO EN INTELIGENCIA ARTIFICIAL A framework for ontology-based library data generation, access and exploitation Doctoral Dissertation of: Daniel Vila-Suero Advisors: Prof. Asunción Gómez-Pérez Dr. Jorge Gracia 2 i To Adelina, Gustavo, Pablo and Amélie Madrid, July 2016 ii Abstract Historically, libraries have been responsible for storing, preserving, cata- loguing and making available to the public large collections of information re- sources. In order to classify and organize these collections, the library commu- nity has developed several standards for the production, storage and communica- tion of data describing different aspects of library knowledge assets. However, as we will argue in this thesis, most of the current practices and standards available are limited in their ability to integrate library data within the largest information network ever created: the World Wide Web (WWW). This thesis aims at providing theoretical foundations and technical solutions to tackle some of the challenges in bridging the gap between these two areas: library science and technologies, and the Web of Data. The investigation of these aspects has been tackled with a combination of theoretical, technological and empirical approaches. Moreover, the research presented in this thesis has been largely applied and deployed to sustain a large online data service of the National Library of Spain: datos.bne.es. Specifically, this thesis proposes and eval- uates several constructs, languages, models and methods with the objective of transforming and publishing library catalogue data using semantic technologies and ontologies. In this thesis, we introduce marimba-framework, an ontology- based library data framework, that encompasses these constructs, languages, mod- els and methods.
    [Show full text]
  • Dealing with Granularity of Time in Temporal Databases
    DEALING WITH GRANULARITY OF TIME IN TEMPORAL DATABASES Gio Wiederhold t Department of Computer Science Stanford University, Stanford, CA 94305-2140, U.S.A. Sushil Jajodia Department of Information Systems and Systems Engineering George Mason University, Fairfax, VA 22030-4444, U.S.A. Witold Litwin* I.N.R.I.A. 78153 Le Chesnay, France ABSTRACT A question that always arises when dealing with temporal informa- tion is the granularity of the values in the domain type. Many different approaches have been proposed; however, the community has not yet come to a basic agree- ment. Most published temporal representations simplify the issue which leads to difficulties in practical applications. In this paper, we resolve the issue of temporal representation by requiring two domain types (event times and intervals), formal- ize useful temporal semantics, and extend the relational operations in such a way that temporal extensions fit into a relational representation. Under these considera- tions, a database system that deals with temporal data can not only present con- sistent temporal semantics to users but perform consistent computational sequences on temporal data from diverse sources. 1. Introduction Large databases do not just collect data about the current state of objects but retain infor- marion about past states as well. In the past, when storage was costly, such data was often not retained online. Aggregations of past information might have been kept, and detail data was archived. Certain applications always required that an adequate history be kept. In medical data- base systems a record of past events is necessary to avoid repeating ineffective or inappropriate treatments, and in other applications legal or planning requirements make keeping of a history desirable.
    [Show full text]
  • A Centralized Ledger Database for Universal Audit and Verification
    LedgerDB: A Centralized Ledger Database for Universal Audit and Verification Xinying Yangy, Yuan Zhangy, Sheng Wangx, Benquan Yuy, Feifei Lix, Yize Liy, Wenyuan Yany yAnt Financial Services Group xAlibaba Group fxinying.yang,yuenzhang.zy,sh.wang,benquan.ybq,lifeifei,yize.lyz,[email protected] ABSTRACT certain consensus protocol (e.g., PoW [32], PBFT [14], Hon- The emergence of Blockchain has attracted widespread at- eyBadgerBFT [28]). Decentralization is a fundamental basis tention. However, we observe that in practice, many ap- for blockchain systems, including both permissionless (e.g., plications on permissioned blockchains do not benefit from Bitcoin, Ethereum [21]) and permissioned (e.g., Hyperledger the decentralized architecture. When decentralized architec- Fabric [6], Corda [11], Quorum [31]) systems. ture is used but not required, system performance is often A permissionless blockchain usually offers its cryptocur- restricted, resulting in low throughput, high latency, and rency to incentivize participants, which benefits from the significant storage overhead. Hence, we propose LedgerDB decentralized ecosystem. However, in permissioned block- on Alibaba Cloud, which is a centralized ledger database chains, it has not been shown that the decentralized archi- with tamper-evidence and non-repudiation features similar tecture is indispensable, although they have been adopted to blockchain, and provides strong auditability. LedgerDB in many scenarios (such as IP protection, supply chain, and has much higher throughput compared to blockchains. It merchandise provenance). Interestingly, many applications offers stronger auditability by adopting a TSA two-way peg deploy all their blockchain nodes on a BaaS (Blockchain- protocol, which prevents malicious behaviors from both users as-a-Service) environment maintained by a single service and service providers.
    [Show full text]
  • Time, Points and Space - Towards a Better Analysis of Wildlife Data in GIS
    Time, Points and Space - Towards a Better Analysis of Wildlife Data in GIS Dissertation zur Erlangung der naturwissenschaftlichen Doktorw¨urde (Dr. sc. nat.) vorgelegt der Mathematisch-naturwissenschaftlichen Fakult¨at der Universit¨at Z¨urich von Stephan Imfeld von Lungern OW Begutachtet von Prof. Dr. Kurt Brassel Prof. Dr. Bernhard Nievergelt Dr. Britta Allg¨ower Prof. Dr. Peter Fisher Z¨urich 2000 Die vorliegende Arbeit wurde von der Mathematisch-naturwissenschaftlichen Fakult¨at der Universit¨at Z¨urich auf Antrag von Prof. Dr. Kurt Brassel und Prof. Dr. Robert Weibel als Dissertation angenommen. How to Catch Running Animals with GIS and See What They’re up to i Abstract Geographical Information Systems are powerful instruments to analyse spatial data. Wildlife researchers and managers are always confronted with spatial data analysis and make use of these systems for various tasks. One important characteristic of the animals unter investigation is their locomotion. Thus the temporal aspects are important, but unfortunately GIS are almost ignorant concerning the analysis of the temporal domain. This thesis is trying to provide a new perspective on how to analyse moving point objects within GIS. A conceptual shift is performed from a space centered view to a way of analysing spatial and temporal aspects in an equally balanced way. For this purpose the family of analytical Time Plots was developed. They represent a completely new approach of how to analyse moving point objects. They transform the data originating from an animal’s movements into a repre- sentation with two time axes and one spatial axis that allows for an effective recognition of spatial patterns within the data.
    [Show full text]
  • SQL and Temporal Database Research: Unified Review and Future Directions
    International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 04 Issue: 09 | Sep -2017 www.irjet.net p-ISSN: 2395-0072 SQL and Temporal Database Research: Unified Review and Future Directions Rose-Mary Owusuaa Mensah1, Vincent Amankona2 1Postgraduate Student, Kwame Nkrumah University of Science and Technology, Kumasi, Ghana 2 Postgraduate Student, Kwame Nkrumah University of Science and Technology, Kumasi, Ghana ---------------------------------------------------------------------***--------------------------------------------------------------------- Abstract - Several attempts to incorporate temporal In the early 1990s, a renowned researcher in the database extensions into the Structured Query Language, SQL, one of community, Richard Snodgrass, proposed that temporal the most popular query languages for databases date back extensions to SQL be developed because few of them to the nineteenth and twentieth century. Although a lot of existed. In response to his proposal [1], a committee was work and research has been done on temporal databases formed to consolidate past research and suggestions from and SQL, there exist very limited literature clearly outlining the research community in order to design temporal the various events which have taken place with regards to extensions to the 1992 edition of the SQL standard. Those temporal extensions of SQL over the years till the present extensions, known as TSQL2, were then developed by this state in a concise document. Consequently, researchers need committee and in 1993, they presented proposals to the to gather several pieces of literature before they can obtain ANSI SQL Technical Committee. Based on responses to the a vivid pictorial timeline of the history and the current state proposals, changes were made to the Language, and the of these temporal extensions for research and software definitive version of the TSQL2 Language Specification development purposes.
    [Show full text]