Lecture @DHBW: Data Warehouse 04 Modeling Andreas Buckenhofer

Total Page:16

File Type:pdf, Size:1020Kb

Lecture @DHBW: Data Warehouse 04 Modeling Andreas Buckenhofer Lecture @DHBW: Data Warehouse 04 Modeling Andreas Buckenhofer Ein Unternehmen der Daimler AG Andreas Buckenhofer Contact/Connect Senior DB Professional xing Since 2009 at Daimler TSS Department: Machine Learning Solutions DOAG DHBW Business Unit: Analytics vcard • Over 20 years experience with • Oracle ACE Associate database technologies • DOAG responsible for InMemory DB • Over 20 years experience with Data • Lecturer at DHBW Daimler TSS GmbH WilhelmWarehousing-Runge-Straße 11, 89081 Ulm / Telefon +49 731 505-06 / Fax +49 731 505• -65 Certified99 Data Vault Practitioner 2.0 [email protected]• International / Internet: www.d projectaimler-tss.com experience • Certified Oracle Professional Sitz und Registergericht: Ulm / HRB-Nr.: 3844 / Geschäftsführung: Martin Haselbach (Vorsitzender), Steffen Bäuerle • Certified IBM Big Data Architect © Daimler TSS I Template Revision Change Log Date Changes 08.03.2020 Initial version Daimler TSS Data Warehouse / DHBW 3 What you will learn today • Understand data modeling (logical models) and physical implementation (physical models) • Explain models for a DWH • Dimensional model • Data Vault model • Understand dimensions and facts • Understand ROLAP & MOLAP Daimler TSS Data Warehouse / DHBW 4 Structuring Data Source: https://www.linkedin.com/feed/update/urn:li:activity:6507927816082845696/ Daimler TSS Data Warehouse / DHBW 5 Conceptual – logical – physical level Different levels of abstraction: • Conceptual (domain) model • Focus on (main) entities and its business definitions! • No attributes • Logical data model • Relational data model (independent of a DBMS or technology) • Logic can't affect performance = no performance optimization on this level • Physical implementation • Representation of a data design for a specific DBMS • RDBMS are the closest to physical independance Daimler TSS Data Warehouse / DHBW 6 Physical implementation in OLTP applications Requirements • Efficient update and delete operations • Efficient read operations • Avoid contradiction in the data – don’t store data twice or multiple times • Easy maintenance of the data model • As little redundancy as possible in the data model Daimler TSS Data Warehouse / DHBW 7 Codd‘s normal forms for DB relations: 1NF • First Normal Form (1NF): • A relation is in first normal form if • the domain of each attribute contains only atomic (simple, indivisible) values. • the value of any attribute in a tuple/row must be a single value from the domain of that attribute, i.e. no attribute values can be sets Daimler TSS Data Warehouse / DHBW 8 Codd‘s normal forms for DB relations: 1NF CD_ID Album Founded Titels 11 Anastacia – Not that kind 1999 1. Not that kind, 2. I‘m outta love, 3 Cowboys & Kisses 12 Pink Floyd – Wish you were here 1964 1. Shine on you crazy diamond 13 Anastacia – Freak of Nature 1999 1. Paid my dues CD_ID Album Performer Founded Track Titels 11 Not that kind Anastacia 1999 1 Not that kind 11 Not that kind Anastacia 1999 2 I‘m outta love 11 Not that kind Anastacia 1999 3 Cowboys & Kisses 12 Wish you were here Pink Floyd 1964 1 Shine on you crazy diamond 13 Freak of Nature Anastacia 1999 1 Paid my dues Daimler TSS Data Warehouse / DHBW 9 Codd‘s normal forms for DB relations: 2NF • Second Normal Form (2NF): • In 1st normal form • Every non-key attribute is fully dependent on the key. There are no dependencies between a partial key and a non-key field. Daimler TSS Data Warehouse / DHBW 10 Codd‘s normal forms for DB relations: 2NF CD_ID Album Performer Founded Track Titels 11 Not that kind Anastacia 1999 1 Not that kind 11 Not that kind Anastacia 1999 2 I‘m outta love 11 Not that kind Anastacia 1999 3 Cowboys & Kisses 12 Wish you were here Pink Floyd 1964 1 Shine on you crazy diamond 13 Freak of Nature Anastacia 1999 1 Paid my dues CD_ID Track Titels 11 1 Not that kind CD_ID Album Performer Founded 11 2 I‘m outta love 11 Not that kind Anastacia 1999 11 3 Cowboys & Kisses 12 Wish you were here Pink Floyd 1964 12 1 Shine on you crazy diamond 13 Freak of Nature Anastacia 1999 13 1 Paid my dues Daimler TSS Data Warehouse / DHBW 11 Codd‘s normal forms for DB relations: 3NF • Third Normal Form (3FN): • In 2nd normal form • No functional dependencies between non key fields: a non-key attribute is dependent from a PK only Codd‘s normal forms for DB relations: 3NF CD_ID Album Performer Founded CD_ID Track Titels 11 Not that kind Anastacia 1999 11 1 Not that kind 12 Wish you were here Pink Floyd 1964 11 2 I‘m outta love 13 Freak of Nature Anastacia 1999 11 3 Cowboys & Kisses 12 1 Shine on you crazy diamond 13 1 Paid my dues CD_ID Album Performer 11 Not that kind Anastacia Performer Founded 12 Wish you were here Pink Floyd Anastacia 1999 13 Freak of Nature Anastacia Pink Floyd 1964 Daimler TSS Data Warehouse / DHBW 13 Codd‘s normal forms - Summary from 1NF to 3NF CD_ID Album Founded Titels 11 Anastacia – Not that kind 1999 1. Not that kind, 2. I‘m outta love, 3 Cowboys & Kisses 12 Pink Floyd – Wish you were here 1964 1. Shine on you crazy diamond 13 Anastacia – Freak of Nature 1999 1. Paid my dues CD_ID Album Performer 11 Not that kind Anastacia CD_ID Track Titels 12 Wish you were here Pink Floyd 11 1 Not that kind 13 Freak of Nature Anastacia 11 2 I‘m outta love n n 11 3 Cowboys & Kisses Performer Founded 12 1 Shine on you crazy diamond Anastacia 1999 13 1 Paid my dues Daimler TSS Pink Floyd 1964 Data Warehouse / DHBW 14 Why database design? “Data modeling is the process of learning about the data, and regardless of technology, this process must be performed for a successful application.” • Learn about the data and promote collective data understanding • Derive security classification and measures • Design for performance • Accelerate development • Improve Software quality • Reduce maintenance costs • Generate code • NoSQL Schema-on-read: understand model versions after years Source quote: Steve Hoberman: Data Modeling for Mongo DB, Technics Publications 2014 Daimler TSS Data Warehouse / DHBW 15 Importance of a good database design Daimler TSS Data Warehouse / DHBW 16 Measuring the quality of a data model data model scorecard Source: Steve Hoberman - Data Modeling Scorecard, Technics Publication 2015 Daimler TSS Data Warehouse / DHBW 17 Exercise: use OLTP physical implementation for DWH? The diagram shows a typical OLTP data model • Customers and products have unique ids and some descriptive attributes • A customer can place an order on a specific date • The order contains one or more products Daimler TSS Data Warehouse / DHBW 18 Exercise: use OLTP physical implementation for DWH? Now consider DWH requirements like non-volatile and time-variant data • Customer Bush marries and takes her husband’s last name • Product number 5 gets a price increase How would you solve such requirements in a data model for the Core Warehouse Layer? Daimler TSS Data Warehouse / DHBW 19 Exercise: use OLTP physical implementation for DWH? Possible solutions: • Add timestamp column as part of the primary key • For all tables, not only for specific tables (e.g. product, customer) • Composite keys can become inefficient and impractical • New tables with head and version data to avoid redundancy • Head table contains static data that does not change (e.g. customer id, birthdate) • Version table contains data that changes (e.g. last name, comments) • Store every change in log tables • Querying tables can become difficult and slow if history is required ("main" table + log tables) Daimler TSS Data Warehouse / DHBW 20 Bad physical implementations Create a SQL statement for: How many "Order Transactions" have been created by "Person/Organisation"? Source: Corr / Stagnitto: Agile Data Warehouse Design, DecisionOne Press, 2011, page 5 Daimler TSS Data Warehouse / DHBW 21 Disadvantages of a direct physical implementation of 3NF for DWH • 3NF is inefficient for query processing • 3NF models are difficult to understand • 3NF gets even more complicated with history added • 3NF not suited for „new“ data sources (JSON, NoSQL, etc.) • DWH needs own data modeling approaches for the Core Warehouse Layer and the Mart Layer Daimler TSS Data Warehouse / DHBW 22 What are candidates for primary keys? Primary keys Natural keys, Sequences, Hash keys Natural Keys Generated Keys Hash Keys „intelligent“ keys that System-generated, (composite) Natural key have a meaning to the unique values, e.g. run through a hash business user sequences (increments) function, e.g. 1, 2, 3, 4, 5, etc. Md5(VIN) VIN (vehicle identifier) GUIDs (globally unique ISO country codes, e.g. identifiers) contain e.g. DE, US, UK MAC address + timestamp to make an identifier unique. Daimler TSS Data Warehouse / DHBW 23 Natural keys Advantages Disadvantages Have a meaning: can be considered as master Varying length (can be short or very long) keys Same value across (OLTP) systems: valid across Meaning can change over time, e.g. VIN business processes standards changed Allow parallel loads in a DWH or Big Data system Can be composite (several fields) which would make joins slower and more complex [concatenation would be possible] Often sequence-driven in OLTP systems (e.g. customer number; collisions possible when integration into DWH is done) Daimler TSS Data Warehouse / DHBW 24 Generated keys Advantages Disadvantages Small byte size (sequences): less storage and Insert performance can be slow (hot spot on faster joins index) Always unique No business meaning Good B*Tree index clustering Data load into DWH cause lookups (Big Data systems often have no sequences but would fail performance-wise with *sequential* sequence generation) Daimler TSS Data Warehouse / DHBW 25 Hash keys Advantages Disadvantages Allow parallel loads in a DWH or Big Data system Computed value can be longer compared to natural keys Ability to join across platforms (e.g.
Recommended publications
  • Sixth Normal Form
    3 I January 2015 www.ijraset.com Volume 3 Issue I, January 2015 ISSN: 2321-9653 International Journal for Research in Applied Science & Engineering Technology (IJRASET) Sixth Normal Form Neha1, Sanjeev Kumar2 1M.Tech, 2Assistant Professor, Department of CSE, Shri Balwant College of Engineering &Technology, DCRUST University Abstract – Sixth Normal Form (6NF) is a term used in relational database theory by Christopher Date to describe databases which decompose relational variables to irreducible elements. While this form may be unimportant for non-temporal data, it is certainly important when maintaining data containing temporal variables of a point-in-time or interval nature. With the advent of Data Warehousing 2.0 (DW 2.0), there is now an increased emphasis on using fully-temporalized databases in the context of data warehousing, in particular with next generation approaches such as Anchor Modeling . In this paper, we will explore the concepts of temporal data, 6NF conceptual database models, and their relationship with DW 2.0. Further, we will also evaluate Anchor Modeling as a conceptual design method in which to capture temporal data. Using these concepts, we will indicate a path forward for evaluating a possible translation of 6NF-compliant data into an eXtensible Markup Language (XML) Schema for the purpose of describing and presenting such data to disparate systems in a structured format suitable for data exchange. Keywords :, 6NF,SQL,DKNF,XML,Semantic data change, Valid Time, Transaction Time, DFM I. INTRODUCTION Normalization is the process of restructuring the logical data model of a database to eliminate redundancy, organize data efficiently and reduce repeating data and to reduce the potential for anomalies during data operations.
    [Show full text]
  • The Design of Multidimensional Data Model Using Principles of the Anchor Data Modeling: an Assessment of Experimental Approach Based on Query Execution Performance
    WSEAS TRANSACTIONS on COMPUTERS Radek Němec, František Zapletal The Design of Multidimensional Data Model Using Principles of the Anchor Data Modeling: An Assessment of Experimental Approach Based on Query Execution Performance RADEK NĚMEC, FRANTIŠEK ZAPLETAL Department of Systems Engineering Faculty of Economics, VŠB - Technical University of Ostrava Sokolská třída 33, 701 21 Ostrava CZECH REPUBLIC [email protected], [email protected] Abstract: - The decision making processes need to reflect changes in the business world in a multidimensional way. This includes also similar way of viewing the data for carrying out key decisions that ensure competitiveness of the business. In this paper we focus on the Business Intelligence system as a main toolset that helps in carrying out complex decisions and which requires multidimensional view of data for this purpose. We propose a novel experimental approach to the design a multidimensional data model that uses principles of the anchor modeling technique. The proposed approach is expected to bring several benefits like better query execution performance, better support for temporal querying and several others. We provide assessment of this approach mainly from the query execution performance perspective in this paper. The emphasis is placed on the assessment of this technique as a potential innovative approach for the field of the data warehousing with some implicit principles that could make the process of the design, implementation and maintenance of the data warehouse more effective. The query performance testing was performed in the row-oriented database environment using a sample of 10 star queries executed in the environment of 10 sample multidimensional data models.
    [Show full text]
  • Translating Data Between Xml Schema and 6Nf Conceptual Models
    Georgia Southern University Digital Commons@Georgia Southern Electronic Theses and Dissertations Graduate Studies, Jack N. Averitt College of Spring 2012 Translating Data Between Xml Schema and 6Nf Conceptual Models Curtis Maitland Knowles Follow this and additional works at: https://digitalcommons.georgiasouthern.edu/etd Recommended Citation Knowles, Curtis Maitland, "Translating Data Between Xml Schema and 6Nf Conceptual Models" (2012). Electronic Theses and Dissertations. 688. https://digitalcommons.georgiasouthern.edu/etd/688 This thesis (open access) is brought to you for free and open access by the Graduate Studies, Jack N. Averitt College of at Digital Commons@Georgia Southern. It has been accepted for inclusion in Electronic Theses and Dissertations by an authorized administrator of Digital Commons@Georgia Southern. For more information, please contact [email protected]. 1 TRANSLATING DATA BETWEEN XML SCHEMA AND 6NF CONCEPTUAL MODELS by CURTIS M. KNOWLES (Under the Direction of Vladan Jovanovic) ABSTRACT Sixth Normal Form (6NF) is a term used in relational database theory by Christopher Date to describe databases which decompose relational variables to irreducible elements. While this form may be unimportant for non-temporal data, it is certainly important for data containing temporal variables of a point-in-time or interval nature. With the advent of Data Warehousing 2.0 (DW 2.0), there is now an increased emphasis on using fully-temporalized databases in the context of data warehousing, in particular with approaches such as the Anchor Model and Data Vault. In this work, we will explore the concepts of temporal data, 6NF conceptual database models, and their relationship with DW 2.0. Further, we will evaluate the Anchor Model and Data Vault as design methods in which to capture temporal data.
    [Show full text]
  • Assessment of Helical Anchors Bearing Capacity for Offshore Aquaculture Applications
    The University of Maine DigitalCommons@UMaine Electronic Theses and Dissertations Fogler Library Summer 8-23-2019 Assessment of Helical Anchors Bearing Capacity for Offshore Aquaculture Applications Leon D. Cortes Garcia University of Maine, [email protected] Follow this and additional works at: https://digitalcommons.library.umaine.edu/etd Recommended Citation Cortes Garcia, Leon D., "Assessment of Helical Anchors Bearing Capacity for Offshore Aquaculture Applications" (2019). Electronic Theses and Dissertations. 3094. https://digitalcommons.library.umaine.edu/etd/3094 This Open-Access Thesis is brought to you for free and open access by DigitalCommons@UMaine. It has been accepted for inclusion in Electronic Theses and Dissertations by an authorized administrator of DigitalCommons@UMaine. For more information, please contact [email protected]. ASSESSMENT OF HELICAL ANCHORS BEARING CAPACITY FOR OFFSHORE AQUACULTURE APPLICATIONS By Leon D. Cortes Garcia B.S. National University of Colombia, 2016 A THESIS Submitted in Partial Fulfillment of the Requirements for the Degree of Master of Science (in Civil Engineering) The Graduate School The University of Maine August 2019 Advisory Committee: Aaron P. Gallant, Professor of Civil Engineering, Co‐Advisor Melissa E. Landon, Professor of Civil Engineering, Co‐Advisor Kimberly D. Huguenard, Professor of Civil Engineering Copyright 2019 Leon D. Cortes Garcia All Rights Reserved ii ASSESSMENT OF HELICAL ANCHORS BEARING CAPACITY FOR OFFSHORE AQUACULTURE APPLICATIONS. By Leon Cortes Garcia Thesis Advisor(s): Dr. Aaron Gallant, Dr. Melissa Landon An Abstract of the Thesis Presented in Partial Fulfillment of the Requirements for the Degree of Master of Science (in Civil Engineering) August 2019 Aquaculture in Maine is an important industry with expected growth in the coming years to provide food in an ecological and environmentally sustainable way.
    [Show full text]
  • Sql Query Optimization for Highly Normalized Big Data
    DECISION MAKING AND BUSINESS INTELLIGENCE SQL QUERY OPTIMIZATION FOR HIGHLY NORMALIZED BIG DATA Nikolay I. GOLOV Lecturer, Department of Business Analytics, School of Business Informatics, Faculty of Business and Management, National Research University Higher School of Economics Address: 20, Myasnitskaya Street, Moscow, 101000, Russian Federation E-mail: [email protected] Lars RONNBACK Lecturer, Department of Computer Science, Stocholm University Address: SE-106 91 Stockholm, Sweden E-mail: [email protected] This paper describes an approach for fast ad-hoc analysis of Big Data inside a relational data model. The approach strives to achieve maximal utilization of highly normalized temporary tables through the merge join algorithm. It is designed for the Anchor modeling technique, which requires a very high level of table normalization. Anchor modeling is a novel data warehouse modeling technique, designed for classical databases and adapted by the authors of the article for Big Data environment and a massively parallel processing (MPP) database. Anchor modeling provides flexibility and high speed of data loading, where the presented approach adds support for fast ad-hoc analysis of Big Data sets (tens of terabytes). Different approaches to query plan optimization are described and estimated, for row-based and column- based databases. Theoretical estimations and results of real data experiments carried out in a column-based MPP environment (HP Vertica) are presented and compared. The results show that the approach is particularly favorable when the available RAM resources are scarce, so that a switch is made from pure in-memory processing to spilling over from hard disk, while executing ad-hoc queries.
    [Show full text]
  • Methods Used in Tieback Wall Design and Construction to Prevent Local Anchor Failure, Progressive Anchorage Failure, and Ground Mass Stability Failure Ralph W
    ERDC/ITL TR-02-11 Innovations for Navigation Projects Research Program Methods Used in Tieback Wall Design and Construction to Prevent Local Anchor Failure, Progressive Anchorage Failure, and Ground Mass Stability Failure Ralph W. Strom and Robert M. Ebeling December 2002 echnology Laboratory Information T Approved for public release; distribution is unlimited. Innovations for Navigation Projects ERDC/ITL TR-02-11 Research Program December 2002 Methods Used in Tieback Wall Design and Construction to Prevent Local Anchor Failure, Progressive Anchorage Failure, and Ground Mass Stability Failure Ralph W. Strom 9474 SE Carnaby Way Portland, OR 97266 Concord, MA 01742-2751 Robert M. Ebeling Information Technology Laboratory U.S. Army Engineer Research and Development Center 3909 Halls Ferry Road Vicksburg, MS 39180-6199 Final report Approved for public release; distribution is unlimited. Prepared for U.S. Army Corps of Engineers Washington, DC 20314-1000 Under INP Work Unit 33272 ABSTRACT: A local failure that spreads throughout a tieback wall system can result in progressive collapse. The risk of progressive collapse of tieback wall systems is inherently low because of the capacity of the soil to arch and redistribute loads to adjacent ground anchors. The current practice of the U.S. Army Corps of Engineers is to design tieback walls and ground anchorage systems with sufficient strength to prevent failure due to the loss of a single ground anchor. Results of this investigation indicate that the risk of progressive collapse can be reduced
    [Show full text]
  • Data Model Matrix
    Delivery Source Validation Central Facts Generic Data Access Tool Access Data Model Matrix Layer: abstraction Data logistics Notation Information (Target) Generic Specific Concern Data Delivery Logical External Facts Enhancement Target Data Mart/ & Access Approach (Source) Integration models Information Levels of Concerns Concerns Artefacts abstraction Validation model models Perspectives Analysis Tool representation Theory Systems Realities § Source System/ § Data exchange § Source System or data § Collection of all UoD’s § Data Integration § Plausibility § Representation § (Temporal) filtering and § Queryable --------------- platform § Data format delivery validation § Structure stability § Data correlation § Cross model validation (Access) selection § Navigation Data definition Data delivery process / § Formal delivery § Validation against logical model § Fact uniformity § Data homogenization § Data enrichment & § Manipulation (CRUD) § Language § Aggregation specification § Data delivery § Data standardization § Minimally constrained§ Unification of concepts and context correlation § Integrity § Type transformation § Tooling Execution level concerns data layer concerns abstraction (for transformation) Temporally flexible § Source model IST to SOLL (Constraints) § Data Application § Performance § Data representation (model repair) § Derivation/ Processing Interface § Scalability § Data packaging transformation (DAPI) - Informal definition - Comprehension § Legal Reports, Dossiers Informal language Model Canvas Narrative Compendium Data
    [Show full text]
  • Anchor Modeling TDWI Presentation
    1 _tÜá e≠ÇÇuùv~ Anchor Modeling I N T H E D A T A W A R E H O U S E Did you know that some of the earliest prerequisites for data warehousing were set over 2500 years ago. It is called an anchor model since the anchors tie down a number of attributes (see picture above). All EER-diagrams have been made with Graphviz. All cats are drawn by the author, Lars Rönnbäck. 1 22 You can never step into the same river twice. The great greek philosopher Heraclitus said ”You can never step into the same river twice”. What he meant by that is that everything is changing. The next time you step into the river other waters are flowing by. Likewise the environment surrounding a data warehouse is in constant change and whenever you revisit them you have to adapt to these changes. Image painted by Henrik ter Bruggen courtsey of Wikipedia Commons (public domain). 2 3 Five Essential Criteria • A future-proof data warehouse must at least fulfill: – Value – Maintainability – Usability – Performance – Flexibility Fail in one and there will be consequences Value – most important, even a very poorly designed data warehouse can survive as long as it is providing good business value. If you fail in providing value, the warehouse will be viewed as a money sink and may be cancelled altogether. Maintainability – you should be able to answer the question: How is the warehouse feeling today? Could be healthy, could be ill! Detect trends, e g in loading times. Usability – must be simple and accessible to the end users.
    [Show full text]
  • Anchor Modelling
    Anchor modeling Beter dan Data Vault? Dit artikel is oorspronkelijk gepubliceerd in Database Magazine 8 / 2009, met andere lay-out en titel. Het copyright voor die versie berust dus bij Array Publishers. Verzoeken voor overname van dit materiaal gelieve men dus aan Array te richten. Grundsätzlich IT M: 06-41966000 W: www.grundsatzlich-it.nl E: [email protected] R. Kunenborg Versie 1.31 / 26 augustus 2009 / Final Contents Anchor Modeling ..................................................................................................................................... 3 Inleiding ............................................................................................................................................... 3 Geschiedenis ....................................................................................................................................... 3 Anchor Modeling ................................................................................................................................. 3 De zesde normaalvorm (6NF) .......................................................................................................... 3 Anchor Modeling ............................................................................................................................. 4 Sleutelmanagement ........................................................................................................................ 7 Views en functies ............................................................................................................................
    [Show full text]
  • Agile Information Modeling in Evolving Data Environments
    Anchor Modeling { Agile Information Modeling in Evolving Data Environments L. R¨onnb¨ack Resight, Kungsgatan 66, 101 30 Stockholm, Sweden O. Regardt Teracom, Kakn¨astornet,102 52 Stockholm, Sweden M. Bergholtz, P. Johannesson, P. Wohed DSV, Stockholm University, Forum 100, 164 40 Kista, Sweden Abstract Maintaining and evolving data warehouses is a complex, error prone, and time consuming activity. The main reason for this state of affairs is that the environment of a data warehouse is in constant change, while the warehouse itself needs to provide a stable and consistent interface to information spanning extended periods of time. In this article, we propose an agile information modeling technique, called Anchor Modeling, that offers non-destructive extensibility mechanisms, thereby enabling robust and flexible management of changes. A key benefit of Anchor Modeling is that changes in a data warehouse environment only require extensions, not modifications, to the data warehouse. Such changes, therefore, do not require immediate modifications of existing applications, since all previous versions of the database schema are available as subsets of the current schema. Anchor Modeling decouples the evolution and application of a database, which when building a data warehouse enables shrinking of the initial project scope. While data models were previously made to capture every facet of a domain in a single phase of development, in Anchor Modeling fragments can be iteratively modeled and applied. We provide a formal and technology independent definition of anchor models and show how anchor models can be realized as relational databases together with examples of schema evolution. We also investigate performance through a number of lab experiments, which indicate that under certain conditions anchor databases perform substantially better than databases constructed using traditional modeling techniques.
    [Show full text]
  • Relational Model of Temporal Data Based on 6Th Normal Form
    D. Golec i dr. Relacijski model vremenskih podataka zasnovan na 6. normalnoj formi ISSN 1330-3651 (Print), ISSN 1848-6339 (Online) https://doi.org/10.17559/TV-20160425214024 RELATIONAL MODEL OF TEMPORAL DATA BASED ON 6TH NORMAL FORM Darko Golec, Viljan Mahnič, Tatjana Kovač Original scientific paper This paper brings together two different research areas, i.e. Temporal Data and Relational Modelling. Temporal data is data that represents a state in time while temporal database is a database with built-in support for handling data involving time. Most of temporal systems provide sufficient temporal features, but the relational models are improperly normalized, and modelling approaches are missing or unconvincing. This proposal offers advantages for a temporal database modelling, primarily used in analytics and reporting, where typical queries involve a small subset of attributes and a big amount of records. The paper defines a distinctive logical model, which supports temporal data and consistency, based on vertical decomposition and sixth normal form (6NF). The use of 6NF allows attribute values to change independently of each other, thus preventing redundancy and anomalies. Our proposal is evaluated against other temporal models and super-fast querying is demonstrated, achieved by database join elimination. The paper is intended to help database professionals in practice of temporal modelling. Keywords: logical model; relation; relational modelling; 6th Normal Form; temporal data Relacijski model vremenskih podataka zasnovan na 6. normalnoj formi Izvorni znanstveni članak Ovaj rad povezuje dva različita područja istraživanja, tj. Temporalne podatke i Relacijsko modeliranje. Temporalni podaci su podaci koji predstavljaju stanje u vremenu, a temporalna baza podataka je baza podataka s ugrađenom podrškom za baratanje s podacima koji uključuju vrijeme.
    [Show full text]
  • AM Cheat Sheet
    BUILDING A FUTURE PROOF DATA WAREHOUSE TDWI 2007 AMSTERDAM Anchor Modeling FLEXIBILITY Where did it all come from? surrogate key found in the anchor and The environment surrounding a Background exactly one attribute value. It should data warehouse is in constant contain meta information similar to that Anchor Modeling is built upon two found in the anchor. If, for a given iden- change. Anchor modeling is built techniques both discovered in the 1970’s; tity, the attribute may change over time, on this premise, such that a large the sixth normal form and entity rela- it should also contain historization in- change on the outside will result tionships. In more recent years the sixth formation. In the case of the attribute normal form has been discussed with in a small change within. having a state or a type, it can also hold respect to storing temporal data. Also, foreign keys of knots. entity relationships has evolved into en- INDEPENDENCE hanced entity relationships, which adds The model itself is independent semantic modeling concepts such as How do you relate entities? of business logic. Rules are de- typing. Ties scriptive rather than physical to Relationships between entities are mod- increase the longevity of the What building blocks are used? eled as ties between anchors. A tie will Constituents thereby relate identities to each other. data warehouse. You have the The most common form is to relate two power to decide how the data There are only four different types of anchors, but there is no theoretical limit should be interpreted. tables used in anchor modeling.
    [Show full text]