A Semantic Framework for Integrating and Publishing Linked Data on the Web

Total Page:16

File Type:pdf, Size:1020Kb

A Semantic Framework for Integrating and Publishing Linked Data on the Web A Semantic Framework for Integrating and Publishing Linked Data on the Web by Aparna Padki A Thesis Presented in Partial Fulfillment of the Requirements for the Degree Master of Science Approved June 2016 by the Graduate Supervisory Committee: Srividya Bansal, Chair Ajay Bansal Timothy Lindquist ARIZONA STATE UNIVERSITY August 2016 ABSTRACT Semantic web is the web of data that provides a common framework and technologies for sharing and reusing data in various applications. In semantic web terminology, linked data is the term used to describe a method of exposing and connecting data on the web from different sources. The purpose of linked data and semantic web is to publish data in an open and standard format and to link this data with existing data on the Linked Open Data Cloud. The goal of this thesis to come up with a semantic framework for integrating and publishing linked data on the web. Traditionally integrating data from multiple sources usually involves an Extract-Transform-Load (ETL) framework to generate datasets for analytics and visualization. The thesis proposes introducing a semantic component in the ETL framework to semi-automate the generation and publishing of linked data. In this thesis, various existing ETL tools and data integration techniques have been analyzed and deficiencies have been identified. This thesis proposes a set of requirements for the semantic ETL framework by conducting a manual process to integrate data from various sources such as weather, holidays, airports, flight arrival, departure and delays. The research questions that are addressed are: (i) to what extent can the integration, generation, and publishing of linked data to the cloud using a semantic ETL framework be automated; (ii) does use of semantic technologies produce a richer data model and integrated data. Details of the methodology, data collection, and application that uses the linked data generated are presented. Evaluation is done by comparing traditional data integration approach with semantic ETL approach in terms of effort involved in integration, data model generated and querying the data generated. i DEDICATION This thesis is dedicated to my husband, Srikar Deshmukh for his constant support, motivation and love. I also dedicate this thesis to my family for their blessings and support. ii ACKNOWLEDGMENTS I would like to thank my supervisor, Dr. Srividya Bansal, for her constant support, guidance and motivation during the course of the development of this thesis and through the entire coursework of my Master’s degree. My appreciation and thanks also goes to Dr. Ajay Bansal and Dr. Timothy Lindquist for their support and encouragement throughout my time in Arizona State University. I would also like to acknowledge and thank Jaydeep Chakraborty for his contribution in the evaluation process and Neha Singh for her contribution in researching various traditional ETL tools. iii TABLE OF CONTENTS Page LIST OF TABLES ................................................................................................................... vi LIST OF FIGURES ............................................................................................................... vii CHAPTER 1 INTRODUCTION ………………………………………………………..1 1.1 Motivation ………………………………………………………...1 1.2 Problem Statement………………………………………………...4 1.3 Scope …………………………………………………………..….4 2 BACKGROUND…………………………………………………….........6 2.1 Semantic Technologies……………………………........................6 2.2 Extract Transform Load Frameworks……………………………11 2.3 Linked Data………………………………………………………17 3 RELATED WORK………………………………………………………20 3.1 Data Integration………………………………………………….20 3.2 Linked Data Generation………………………………………….23 3.3 Semantic ETL…………………………………………………....25 4 DATA EXTRACTION…………………………………………………..28 4.1 Flights Delays Data Set………………………………………….29 iv CHAPTER Page 4.2 Weather Data Set…………………………………………..........29 4.3 Holiday Data Set…………………………………………...........30 5 SEMANTIC DATA INTEGRATION – METHODOLOGY ..…………31 5.1 Semantic Data Model……………………………………………31 5.2 Linked Data Integration………………………………………….32 6 CASE STUDY IMPLEMENTATION……...…………………………...35 6.1 High Level Design ………………………………………………35 6.2 Implementation Details…………………………..........................37 7 EVALUATION AND RESULTS…………………………...…………...41 7.1 Requirements For Semantic ETL Framework ...………………...41 7.2 Comparison with Traditional ETL…………….………………...49 8 CONCLUSION AND FUTURE WORK………………………………..55 8.1 Conclusion……………………………………………………….55 8.2 Future Work……………………………………………………...55 REFERENCES....................................................................................................................... 57 APPENDIX A WEB APPLICATION SOURCE CODE ............................................................... 62 B PROOF OF CONCEPT SOURCE CODE ............................................................. 74 v LIST OF TABLES Table Page 1. Comparison of ETL Tools ................................................................................... 15 2. Number of Entities in the Ontology…………………………………………...31 3. Size of Individual Datasets……………………………………………………33 4. Query Execution time Traditional vs Semantic……………………………….52 vi LIST OF FIGURES Figure Page 1. Data warehouse Architecture ........................................................................... 2 2. Statement in Form of a Simple Graph ............................................................. 7 3. Hierarchy in the Ontology ............................................................................... 9 4. Clover ETL Sample Workflow ...................................................................... 12 5. LOD Cloud Diagram ..................................................................................... 18 6. Relationship Between Datasets Used and their Sources ............................... 28 7. Semantic Data Model ..................................................................................... 32 8. Data Model in Karma Tool ............................................................................ 33 9. Landing Page of Web Application ................................................................ 36 10. Result of First Search .................................................................................... 36 11. Architecture for Case Study ......................................................................... 37 12. Semantic ETL Process .................................................................................. 43 13. Landing Page ................................................................................................. 48 14. Page After Uploading Ontology File and Selecting a Class ........................ 48 15. Extending Linked Vocabulary - Architecture ............................................. 49 16. Database Schema .......................................................................................... 52 17. Data integration and linking using Semantic ETL ....................................... 53 18. Data integration and linking using Traditional ETL .................................... 54 vii CHAPTER 1 INTRODUCTION 1.1 Motivation With the widespread usage of internet, social media, technology devices and mobile applications, big data is only getting bigger and the sources of this data, types of format of data (structured and unstructured) are increasing as well. In order to be able to utilize this big data, a data integration process from these aforementioned heterogeneous sources is imperative. A wide variety of data warehouses exist that store these huge datasets to be analyzed and visualized for various business/research needs. The definition of a data warehouse is a central repository for all or significant parts of data that an enterprise’s various business systems collect. From this definition, we can understand two important concepts in a data warehouse: (1) The source(s) of data populated in this repository is usually multiple (2) the repository should hold this large amount of data in a coherent/useful format, irrespective of the disparate sources. The general architecture of a data warehouse is shown in Figure 1. 1 Figure 1 Data warehouse Architecture In order to store data from a wide range of sources into a central repository, Extract- Transform-Load (ETL) tools are widely used [1]. Each stage in an ETL process can be defined as: Extract the data – Obtain data from different sources and convert it into an intermediate format. Transform the data - Business rules are applied to the data. Also, to make the data more efficient columns are split or merged. This is done in order to combine data from the extract phase to make it more meaningful and useful. This phase also involves cleaning of the data. Load the data – The transformed data is loaded to a data warehouse for a single access of this data for analytics and visualizations. 2 In the semantic web world, we can draw a parallel to this process of loading a data warehouse for publishing linked data to the linked open data cloud [2]. A Semantic component (i.e., richer representation of the data model to provide more meaning, context, and relationships) in the transform phase is the major missing component in the traditional ETL tools if it has to be used for generating linked data. By introducing this semantic element, we can associate a domain-specific ontology to the key elements/fields in the dataset that is under integration. This is essential in generating linked data and connecting it to the linked open data (LOD) cloud. Using a semantic ETL framework, a richer
Recommended publications
  • Normalized Form Snowflake Schema
    Normalized Form Snowflake Schema Half-pound and unascertainable Wood never rhubarbs confoundedly when Filbert snore his sloop. Vertebrate or leewardtongue-in-cheek, after Hazel Lennie compartmentalized never shreddings transcendentally, any misreckonings! quite Crystalloiddiverted. Euclid grabbles no yorks adhered The star schemas in this does not have all revenue for this When done use When doing table contains less sensible of rows Snowflake Normalizationde-normalization Dimension tables are in normalized form the fact. Difference between Star Schema & Snow Flake Schema. The major difference between the snowflake and star schema models is slot the dimension tables of the snowflake model may want kept in normalized form to. Typically most of carbon fact tables in this star schema are in the third normal form while dimensional tables are de-normalized second normal. A relation is danger to pause in First Normal Form should each attribute increase the. The model is lazy in single third normal form 1141 Options to Normalize Assume that too are 500000 product dimension rows These products fall under 500. Hottest 'snowflake-schema' Answers Stack Overflow. Learn together is Star Schema Snowflake Schema And the Difference. For step three within the warehouses we tested Redshift Snowflake and Bigquery. On whose other hand snowflake schema is in normalized form. The CWM repository schema is a standalone product that other products can shareeach product owns only. The main difference between in two is normalization. Families of normalized form snowflake schema snowflake. Star and Snowflake Schema in Data line with Examples. Is spread the dimension tables in the snowflake schema are normalized. Like price weight speed and quantitiesie data execute a numerical format.
    [Show full text]
  • Data Warehouse: an Integrated Decision Support Database Whose Content Is Derived from the Various Operational Databases
    1 www.onlineeducation.bharatsevaksamaj.net www.bssskillmission.in DATABASE MANAGEMENT Topic Objective: At the end of this topic student will be able to: Understand the Contrasting basic concepts Understand the Database Server and Database Specified Understand the USER Clause Definition/Overview: Data: Stored representations of objects and events that have meaning and importance in the users environment. Information: Data that have been processed in such a way that they can increase the knowledge of the person who uses it. Metadata: Data that describes the properties or characteristics of end-user data and the context of that data. Database application: An application program (or set of related programs) that is used to perform a series of database activities (create, read, update, and delete) on behalf of database users. WWW.BSSVE.IN Data warehouse: An integrated decision support database whose content is derived from the various operational databases. Constraint: A rule that cannot be violated by database users. Database: An organized collection of logically related data. Entity: A person, place, object, event, or concept in the user environment about which the organization wishes to maintain data. Database management system: A software system that is used to create, maintain, and provide controlled access to user databases. www.bsscommunitycollege.in www.bssnewgeneration.in www.bsslifeskillscollege.in 2 www.onlineeducation.bharatsevaksamaj.net www.bssskillmission.in Data dependence; data independence: With data dependence, data descriptions are included with the application programs that use the data, while with data independence the data descriptions are separated from the application programs. Data warehouse; data mining: A data warehouse is an integrated decision support database, while data mining (described in the topic introduction) is the process of extracting useful information from databases.
    [Show full text]
  • The Design of Multidimensional Data Model Using Principles of the Anchor Data Modeling: an Assessment of Experimental Approach Based on Query Execution Performance
    WSEAS TRANSACTIONS on COMPUTERS Radek Němec, František Zapletal The Design of Multidimensional Data Model Using Principles of the Anchor Data Modeling: An Assessment of Experimental Approach Based on Query Execution Performance RADEK NĚMEC, FRANTIŠEK ZAPLETAL Department of Systems Engineering Faculty of Economics, VŠB - Technical University of Ostrava Sokolská třída 33, 701 21 Ostrava CZECH REPUBLIC [email protected], [email protected] Abstract: - The decision making processes need to reflect changes in the business world in a multidimensional way. This includes also similar way of viewing the data for carrying out key decisions that ensure competitiveness of the business. In this paper we focus on the Business Intelligence system as a main toolset that helps in carrying out complex decisions and which requires multidimensional view of data for this purpose. We propose a novel experimental approach to the design a multidimensional data model that uses principles of the anchor modeling technique. The proposed approach is expected to bring several benefits like better query execution performance, better support for temporal querying and several others. We provide assessment of this approach mainly from the query execution performance perspective in this paper. The emphasis is placed on the assessment of this technique as a potential innovative approach for the field of the data warehousing with some implicit principles that could make the process of the design, implementation and maintenance of the data warehouse more effective. The query performance testing was performed in the row-oriented database environment using a sample of 10 star queries executed in the environment of 10 sample multidimensional data models.
    [Show full text]
  • Chapter 7 Multi Dimensional Data Modeling
    Chapter 7 Multi Dimensional Data Modeling Fundamentals of Business Analytics” Content of this presentation has been taken from Book “Fundamentals of Business Analytics” RN Prasad and Seema Acharya Published by Wiley India Pvt. Ltd. and it will always be the copyright of the authors of the book and publisher only. Basis • You are already familiar with the concepts relating to basics of RDBMS, OLTP, and OLAP, role of ERP in the enterprise as well as “enterprise production environment” for IT deployment. In the previous lectures, you have been explained the concepts - Types of Digital Data, Introduction to OLTP and OLAP, Business Intelligence Basics, and Data Integration . With this background, now its time to move ahead to think about “how data is modelled”. • Just like a circuit diagram is to an electrical engineer, • an assembly diagram is to a mechanical Engineer, and • a blueprint of a building is to a civil engineer • So is the data models/data diagrams for a data architect. • But is “data modelling” only the responsibility of a data architect? The answer is Business Intelligence (BI) application developer today is involved in designing, developing, deploying, supporting, and optimizing storage in the form of data warehouse/data marts. • To be able to play his/her role efficiently, the BI application developer relies heavily on data models/data diagrams to understand the schema structure, the data, the relationships between data, etc. In this lecture, we will learn • About basics of data modelling • How to go about designing a data model at the conceptual and logical levels? • Pros and Cons of the popular modelling techniques such as ER modelling and dimensional modelling Case Study – “TenToTen Retail Stores” • A new range of cosmetic products has been introduced by a leading brand, which TenToTen wants to sell through its various outlets.
    [Show full text]
  • ER/Studio Enterprise Data Modeling
    ER/Studio Enterprise Data Modeling ER/Studio®, a model-driven data architecture and database design solution, helps companies discover, document, and reuse data assets. With round-trip database support, data architects have the power to thoroughly analyze existing data sources as well as design and implement high quality databases that reflect business needs. The highly-readable visual format enhances communication across job functions, from business analysts to application developers. ER/Studio Enterprise also enables team and enterprise collaboration with its repository. • Enhance visibility into your existing data assets • Effectively communicate models across the enterprise Related Products • Improve data consistency • Trace data origins and whereabouts to enhance data integration and accuracy ER/Studio Viewer View, navigate and print ER/Studio ENHANCE VISIBILITY INTO YOUR EXISTING DATA ASSETS models in a view-only environ- ment. As data volumes grow and environments become more complex corporations find it increasingly difficult to leverage their information. ER/Studio provides an easy- Describe™ to-use visual medium to document, understand, and publish information about data assets so that they can be harnessed to support business objectives. Powerful Design, document, and maintain reverse engineering of industry-leading database systems allow a data modeler to enterprise applications written in compare and consolidate common data structures without creating unnecessary Java, C++, and IDL for better code duplication. Using industry standard notations, data modelers can create an infor- quality and shorter time to market. mation hub by importing, analyzing, and repurposing metadata from data sources DT/Studio® such as business intelligence applications, ETL environments, XML documents, An easy-to-use visual medium to and other modeling solutions.
    [Show full text]
  • Advantages of Dimensional Data Modeling
    Advantages of Dimensional Data Modeling 2997 Yarmouth Greenway Drive Madison, WI 53711 (608) 278-9964 www.sys-seminar.com Advantages of Dimensional Data Modeling 1 Top Ten Reasons Why Your Data Model Needs a Makeover 1. Ad hoc queries are difficult to construct for end-users or must go through database “gurus.” 2. Even standard reports require considerable effort and detail knowledge of the database. 3. Data is not integrated or is inconsistent across sources. 4. Changes in data values or in data sources cannot be handled gracefully. 5. The structure of the data does not mirror business processes or business rules. 6. The data model limits which BI tools can be used. 7. There is no system for maintaining change history or collecting metadata. 8. Disk space is wasted on redundant values. 9. Users who might benefit from the data don’t use it. 10.Maintenance is tedious and ad hoc. 2 Advantages of Dimensional Data Modeling Part 1 3 Part 1 - Data Model Overview •What is data modeling and why is it important? •Three common data models: de-normalized (SAS data sets) normalized dimensional model •Benefits of the dimensional model 4 What is data modeling? • The generalized logical relationship among tables • Usually reflected in the physical structure of the tables • Not tied to any particular product or DBMS • A critical design consideration 5 Why is data modeling important? •Allows you to optimize performance •Allows you to minimize costs •Facilitates system documentation and maintenance • The dimensional data model is the foundation of a well designed data mart or data warehouse 6 Common data models Three general data models we will review: De-normalized Expected by many SAS procedures Normalized Often used in transaction based systems such as order entry Dimensional Often used in data warehouse systems and systems subject to ad hoc queries.
    [Show full text]
  • Bio-Ontologies Submission Template
    Relational to RDF mapping using D2R for translational research in neuroscience Rudi Verbeeck*1, Tim Schultz2, Laurent Alquier3 and Susie Stephens4 Johnson & Johnson Pharmaceutical Research and Development 1 Turnhoutseweg 30, Beerse, Belgium; 2 Welch & McKean Roads, Spring House, PA, United States; 3 1000 Route 202, Rari- tan, NJ, United States and 4 145 King of Prussia Road, Radnor, PA, United States ABSTRACT Relational database technology has been developed as an Motivation: To support translational research and external approach for managing and integrating data in a highly innovation, we are evaluating the potential of the semantic available, secure and scalable architecture. With this ap- web to integrate data from discovery research through to the proach, all metadata is embedded or implicit in the applica- clinical environment. This paper describes our experiences tion or metadata schema itself, which results in performant in mapping relational databases to RDF for data sets relating queries. However, this architecture makes it difficult to to neuroscience. share data across a large organization where different data- Implementation: We describe how classes were identified base schemata and applications are being used. in the original data sets and mapped to RDF, and how con- Semantic web offers a promising approach to interconnect nections were made to public ontologies. Special attention databases across an organization, since the technology was was paid to the mapping of experimental measures to RDF designed to function within the distributed environment of and how it was impacted by the relational schemata. the web. Resource Description Framework (RDF) and Web Results: Mapping from relational databases to RDF can Ontology Language (OWL) are the two main semantic web benefit from techniques borrowed from dimensional model- standard recommendations.
    [Show full text]
  • Round-Trip Database Support Gives ER/Studio Data Architect Users The
    Round-trip database support gives ER/Studio Data Architect users the power to easily reverse-engineer, compare and merge, and visually document data assets residing in diverse locations from data centers to mobile platforms. You’ll leverage Enterprise data as corporate asset more effectively and can be rest assured that compliance is supported for business standards and mandatory regulations—essential factors in an organizational data governance program. Automate and scale your data modeling “We had to use a tool like Visio and do everything by hand before. We’re talking about thousands of tables with subsequent relationships. Now ER/Studio automates the most time consuming tasks.” - Jason Soroko, Business Architect at Entrust Uncover your database inconsistencies “Embarcadero tools detected inconsistencies we didn’t even know existed in our systems, saving us from problems down the road.” - U.S. Bancorp Piper Jaffray The Challenge of Fully Leveraging Enterprise Data This all-too-common scenario forces non-compliance with business standards and mandatory regulations, As organizations grow and data proliferates, ad hoc systems for while preventing business executives from the benefit of storing, analyzing, and utilizing that data start to appear and incorporating all essential data in to their decision-making are generally located near to the business unit that needs it. This process. Data management professionals face three distinct practice results in disparately located databases storing different challenges: versions and formats of the same data and an enterprise that will suffer from multiple views and instances of a single data • Reduce duplication and risk associated with multiple data capture and storage environments point.
    [Show full text]
  • Data Model Transformations: Relational to Dimensional
    https://doi.org/10.48009/1_iis_2014_285-291 Issues in Information Systems Volume 15, Issue I, pp. 285-291, 2014 DATA MODEL TRANSFORMATIONS: RELATIONAL TO DIMENSIONAL Vladan Jovanovic, Georgia Southern University, [email protected] James Harris, Georgia Southern University, [email protected] ABSTRACT The paper extends the Conceptual Dimensional Fact Model to support the modeling of data marts by formalizing patterns of graph transformations from a relational data base schema to a data mart schema, and by standardizing all visual representations using the Idef1X standard notation and a single supporting case tool. The proposed modeling and automation approach has the potential to significantly speed up prototyping, improve quality of design documentation, and allow verification by developers and validation by users. Keywords: Dimensional Fact Model, Data Model, Idef1X notation, Relational DB, Graph Transformations, Dimensional Model, and Data Mart. INTRODUCTION Data marts (DM's) are access layers to data warehouses that allow subunits of an organization to efficiently access data in a Data Warehouse. Golfarelli [5] contends that a lack of adequate conceptual modeling and analysis in designing data warehouses and particularly DM's is detrimental to quality, speed and utility. The field of database design not only recognizes the need for separating conceptual design from logical and physical design, but also provides a variety of different models and techniques [2,3,6,7,12]. The review of approaches to conceptual DM design
    [Show full text]
  • Object Triple Mapping Bridging Semantic Web and Object- Oriented Programming
    Object Triple Mapping Bridging Semantic Web and object- oriented programming The use of Semantic Web and ontologies in creating smart applications has drastically increased during the recent years due to the high potential in allowing machines to reason on provided semantics. Providing powerful tools that can improve the development of applications based on ontologies can accelerate the adoption of Semantic Web. For this purpose, many tools such as ontology editors, ontology reasoners, triple store frameworks and object triple mapping systems have been developed. This article introduces the concept of object triple mapping, which essentially consists of a bridge between RDF/OWL and object-oriented programming. What is object-oriented objects that they want to manipulate as well as the relationships between them (process known as programming? data modelling). Object-oriented programming (OOP) is a paradigm for software development, based on the concept of What Semantic Web and OOP objects, which can contain data, in the form of fields (often known as attributes or properties), have in common? and code, in the form of procedures (often known Semantic Web shares a few major characteristics as methods). with OOP: The fundamental characteristics of OOP are: ▪ Semantic Web has classes for defining data concepts; 1. Inheritance – Parent to child relationships; ▪ Semantic Web has instances for actual data 2. Polymorphism – Overloading and overriding values; class members; ▪ Inheritance among classes is supported; 3. Encapsulation – Hiding data behind ▪ Strongly typed data types are supported operations; (string, integer, etc.); 4. Abstraction – Reveal mechanisms that are ▪ Whole/part relationships are supported. only relevant for the use of other objects.
    [Show full text]
  • Query Optimizing for On-Line Analytical Processing Adventures in the Land of Heuristics
    Aalto University School of Science Master's Programme in Computer, Communication and Information Sciences Jonas Berg Query Optimizing for On-line Analytical Processing Adventures in the land of heuristics Master's Thesis Espoo, May 22, 2017 Supervisor: Professor Eljas Soisalon-Soininen Advisors: Jarkko Miettinen M.Sc. (Tech.) Marko Nikula M.Sc. (Tech) Aalto University School of Science Master's Programme in Computer, Communication and In- ABSTRACT OF formation Sciences MASTER'S THESIS Author: Jonas Berg Title: Query Optimizing for On-line Analytical Processing { Adventures in the land of heuristics Date: May 22, 2017 Pages: vii + 71 Major: Computer Science Code: SCI3042 Supervisor: Professor Eljas Soisalon-Soininen Advisors: Jarkko Miettinen M.Sc. (Tech.) Marko Nikula M.Sc. (Tech) Newer database technologies, such as in-memory databases, have largely forgone query optimization. In this thesis, we presented a use case for query optimization for an in-memory column-store database management system used for both on- line analytical processing and on-line transaction processing. To date, the system in question has used a na¨ıve query optimizer for deciding on join order. We went through related literature on the history and evolution of database technology, focusing on query optimization. Based on this, we analyzed the current system and presented improvements for its query processing abilities. We implemented a new query optimizer and experimented with it, seeing how it performed on several queries, concluding that it is a successful improvement
    [Show full text]
  • Data Modeling Immersion 3NF DV & Dimensional
    Course Description Modeling: Operational, Data Warehousing & Data Marts Operational DW DMs GENESEE ACADEMY, LLC 2013 Course Developed by: Hans Hultgren DATA MODELING IMMERSION Modeling: Operational, Data Warehousing & Data Marts Overview Data Modeling is the process of database design. As with most design processes, data modeling is both an art and a science. Art: in that it is a creative process whereby we analyze, design and model specific solutions based on unique requirements. Science: in that we apply modeling approaches, methodologies and specific modeling patterns based on the constraints, variables and performance characteristics of the database we are creating. Our operational systems (core business applications), our data warehouse, and our data marts each have very different constraints, variables and performance characteristics. This course teaches data modeling and covers the main data modeling techniques related to each of these three major architectural layers. In this class you will learn data modeling using the normalized (3NF) modeling approach for operational systems, the ensemble data vault modeling approach for the data warehouse, and the dimensional (Star Schema) modeling approach for data marts. This is a two (2) day course delivered in the classroom. 1/1/2013 | DATA MODELING IMMERSION 1 Course Description This course covers the core principles of data modeling through lectures and hands-on labs and exercises. Providing a solid overview of current techniques for modeling operational systems, data warehouses, and data marts. In covering these areas, the course considers data modeling and design using normalized data modeling 3rd Normal Form for Operational Systems, Ensemble Data Vault modeling for the Data Warehouse, and Dimensional modeling (Star Schema) for Data Marts.
    [Show full text]