Data Mapping: a Key Step to Staying Compliant with Data Security Regulations

Total Page:16

File Type:pdf, Size:1020Kb

Data Mapping: a Key Step to Staying Compliant with Data Security Regulations An iBridge Point of View Data Mapping: A Key Step to Staying Compliant with Data Security Regulations How data mapping helps organizations comply with the GDPR, CCPA, and more. 1 Data security, data integrity, and data management tasks, such as data privacy are key concerns for today’s transformation or mediation between a businesses. In particular, the GDPR in source and a destination. It also the EU and CCPA in California mandate provides database schema mapping for organizations to take the necessary pre-integration data cleansing. measures to protect data at inception, in motion, and at rest. The Importance of Data Mapping Securing your data not only protects Data analytics is a key component in your customers’ information but also any digital transformation (DX) strategy helps you avoid hefty penalties. For in today’s data-driven business instance, GDPR fines are rising environment. Data mapping allows exponentially (up to 10 million euros), companies to have more efficient and CCPA penalties aren’t far behind. document collection to organize their business critical data. Data mapping But how can you be sure that all your also provides easy access to data is fully protected at all times? Are information for timely decision-making. you assuming that your data is protected, or do you know for certain Data mapping has greater accuracy and that everything in your system is consistency, increases the efficiency of airtight? What can you do to take the all business operations, and allows guesswork out of the data leadership to identify emerging trends. management process? Data mapping helps companies The first step to securing your data is minimize potential issues that would knowing where it is, how it travels from lead to lower profits because it allows one location to another, and who has decisions to become more data-driven. access to it. Data mapping takes the Leveraging data mapping for decision- guesswork out by providing a factual making can deliver higher quality account of the locations of all your customer experience to increase critical information so you can ensure profitability. data security while obeying various regulations. What’s Data Mapping? Data mapping is the process of associating data elements between two distinct data models or data fields from a source file to related targeted fields. It synthesizes information from multiple sources and is a critical step for a variety of data integration and 2 Most importantly, data mapping is an element of data integration and enhances various data management data management, such as data tasks that are critical to ensuring data warehousing. security, enforcing data governance and staying compliant Data Warehousing: the aggregation of with industry regulations. data that has already been migrated, integrated, and transformed into one How Data Mapping Helps source for comparison or analysis to Organizations Stay Compliant with improve business intelligence. Data Regulations mapping guarantees that data gets to its intended destination. Data mapping’s role in various data management tasks provides enterprises Here are the key data mapping steps you with an accurate picture of what data should incorporate in your data they hold, where it is, how it’s moved management process: throughout the organization, and what needs protecting. Data Discovery Tracking such information will also help As data is used throughout an you set up the appropriate access organization, it may “leak” into many controls and identify where you should places—for example, in development implement measures like sandboxes, backup files, legacy systems, pseudonymization, encryption, cloud platforms, and standalone anonymization, and aggregation to databases set up by individual ensure compliance with various departments. regulations, such as the GDPR and CCPA. Data Migration: the one-time movement of data from one system to another. Data mapping ensures that the source fields are transferred to the destination fields safely and accurately. Data Integration: the continuing process of integrating data from different sources into a unified view. Data integration enables analytical tools to be able to produce effective, actionable business decisions. Data Transformation: the conversion process of data from one format to a designated format. Data transformation 2 By taking inventory of every database encryption to protect data at rest, in and every instance of it, you’ll have a transit, or in use, as well as setting up full picture of all the data that’s used, role-based access control to allow only accessed, and needs to be protected designated users to view, edit, or within your organization. You can also delete specific information. map out every location and show the direction of data flow to track data Data Management movement within your company. Most regulations, including the GDPR Data Classification and CCPA, are ongoing requirements, so you’ll need to manage your data This is a process of the organization of consistently. You can leverage the data by relevant categories so that it knowledge and insights you gain in the may be secured more efficiently within data mapping process to track data your database system. Having usage and movement, as well as categorized data will allow your maintain records for future audit, database system to protect processing, or decision-making. information with automated technologies. Data classification is You can automate many of these data particularly important when it comes management tasks to increase to risk management, compliance and operational cost-efficiency while data security. It is crucial for reducing errors and delays associated companies to classify data so that with manual processes. Automation there is coverage by the GDPR so it is can also minimize human oversight easily identifiable and has proper that can lead to costly data breaches security precautions. and hefty penalties. Surface Attack Area Reduction Data Mapping Techniques and Software After discovering and classifying your data, you now have a good idea of There are various data mapping what you need to protect and where techniques you can use based on the the security gaps are. Assessing your nature of the information, the cybersecurity risks can help focus your structure of your system, and the IT resources on the vulnerable areas objective of the process. Examples that are most susceptible for attacks include: and breaches. Data-driven mapping, which discovers By identifying where and how your complex mapping between two data data may be exposed to unauthorized sets using heuristics and statistics users, you can take the necessary evaluations. It can automatically find actions to minimize exposure. the transformations between the two Examples include masking personal data types. data for analytics purposes, using 2 An intuitive, cloud-based tool that is Instant data preview: real-time testing designed to have automated repetitive and mapping validation that lets users tasks will save time and decrease the view processed or raw data at any step risk of human error. during the process to prevent mapping errors. How to Select the Right Data Mapping Software “Smart match” functionality: the capability to resolve discrepancies in If you want to streamline your data field names to minimize issues caused mapping process and ensure its by naming conflicts. accuracy, you need to use the right data mapping software. Here are some Data visualization: the ability to key features and criteria to consider: visualize data can help you understand the relationships among various data Support for diverse data sources: the sources and the movement of data. ability to connect to a wide range of structured, unstructured, and semi- Automated data subject requests and structured data sources such as consent management: a feature that databases, web services, REST APIs, allows you to request reports on their and flat file formats. data or manage consent about how their data is used and have the Code-free user interface: easy to use preferences updated in real-time drag-and-drop graphical interface throughout your systems. empowers business users to take charge of the data management Reporting: customizable and process with minimal IT support. synchronal reporting can help you assess, identify, and minimize risks in a Automation and job scheduling: the timely manner while ensuring that you ability to orchestrate a complete generate proper records for audit database mapping workflow with a purposes. time-based and event-triggered job scheduling function. Data Mapping Is Indispensable for Today’s Data-Driven Enterprises Staying compliant with various regulations is non-negotiable for any organization. Without data mapping, which gives you a clear picture of where your data is, how it’s used, and who has access to it, you’re shooting in the dark. As such, data mapping is a key component of any data management system, ensuring the safety and integrity of your organization’s business-critical information. However, setting up a cost-efficient system to manage workflows, comply with various regulations, and ensure that nothing falls through the cracks can be challenging without the right experience and expertise. Here at iBridge, our team of digital transformation experts implements data management solutions for clients in healthcare, legal, manufacturing, utilities, education, finance, and more. We work closely with our clients to understand their business structure so we can set up the best data management systems for their organizations. Get in touch to see how we can help your organization design and implement a successful digital transformation strategy. iBridge is a Digital Transformation Company. We help our clients collect, manage, and analyze their data to create meaningful operational control and improved profitability. For more than a decade, iBridge has successfully distilled complex information into actionable results. www.ibridgellc.com.
Recommended publications
  • Data Mapping: Definition, Tools, and Techniques
    Data Mapping: Definition, Tools, and Techniques Enterprise data is getting more dispersed and voluminous by the day, and at the same time, it has become more important than ever for businesses to leverage data and transform it into actionable insights. However, enterprises today collect information from an array of data points, and they may not always speak the same language. To integrate this data and make sense of it, data mapping is used which is the process of establishing relationships between separate data models. What is Data Mapping? In simple words, data mapping is the process of mapping data fields from a source file to their related target fields. For example, in Figure 1, ‘Name,’ ‘Email,’ and ‘Phone’ fields from an Excel source are mapped to the relevant fields in a Delimited file, which is our destination. Data mapping tasks vary in complexity, depending on the hierarchy of the data being mapped, as well as the disparity between the structure of the source and the target. Every business application, whether on-premise or cloud, uses metadata to explain the data fields and attributes that constitute the data, as well as semantic rules that govern how data is stored within that application or repository. For example, Microsoft Dynamics CRM contains several data objects, such as Leads, Opportunities, and Competitors. Each of these data objects has several fields like Name, Account Owner, City, Country, Job Title, and more. The application also has a defined schema along with attributes, enumerations, and mapping rules. Therefore, if a new record is to be added to the schema of a data object, a data map needs to be created from the data source to the Microsoft Dynamics CRM account.
    [Show full text]
  • Managing Data in Motion This Page Intentionally Left Blank Managing Data in Motion Data Integration Best Practice Techniques and Technologies
    Managing Data in Motion This page intentionally left blank Managing Data in Motion Data Integration Best Practice Techniques and Technologies April Reeve AMSTERDAM • BOSTON • HEIDELBERG • LONDON NEW YORK • OXFORD • PARIS • SAN DIEGO SAN FRANCISCO • SINGAPORE • SYDNEY • TOKYO Morgan Kaufmann is an imprint of Elsevier Acquiring Editor: Andrea Dierna Development Editor: Heather Scherer Project Manager: Mohanambal Natarajan Designer: Russell Purdy Morgan Kaufmann is an imprint of Elsevier 225 Wyman Street, Waltham, MA 02451, USA Copyright r 2013 Elsevier Inc. All rights reserved. No part of this publication may be reproduced or transmitted in any form or by any means, electronic or mechanical, including photocopying, recording, or any information storage and retrieval system, without permission in writing from the publisher. Details on how to seek permission, further information about the Publisher’s permissions policies and our arrangements with organizations such as the Copyright Clearance Center and the Copyright Licensing Agency, can be found at our website: www.elsevier.com/permissions. This book and the individual contributions contained in it are protected under copyright by the Publisher (other than as may be noted herein). Notices Knowledge and best practice in this field are constantly changing. As new research and experience broaden our understanding, changes in research methods or professional practices, may become necessary. Practitioners and researchers must always rely on their own experience and knowledge in evaluating and using any information or methods described herein. In using such information or methods they should be mindful of their own safety and the safety of others, including parties for whom they have a professional responsibility.
    [Show full text]
  • Transforming MARCXML Records Using XSLT
    Transforming MARCXML Records Using XSLT Violeta Ilik Head, Digital Systems & Collection Services Digital Innovations Librarian Galter Health Sciences Library Northwestern University Clinical and Translational Sciences Institute Chicago, IL March 4, 2015 Teun Hocks Violeta Ilik vivoweb.org Hosted by ALCTS, Association for Library Collections and Technical Services + ALCTS - Library Resources & Technical Services n Ilik, V., Storlien, J., and Olivarez, J. (2014). “Metadata Makeover: Transforming MARC Records Using XSLT.” Library Resources and Technical Services 58 (3): 187-208. “The knowledge gained from learning information technology can be used to experiment with methods of transforming one metadata schema into another using various software solutions” Violeta Ilik vivoweb.org Hosted by ALCTS, Association for Library Collections and Technical Services + Why Catalogers? n Survey conducted by Ma in 2007 revealed that the metadata qualifications and responsibilities required knowledge of MARC, crosswalks, XML, OAI … n Analysis of cataloging position description performed by Park, Lu, and Marion reveal that advances in technology have created a new realm of desired skills, qualifications and responsibilities for catalogers. n Terry Reese’s MarcEdit is a best friend to everyone in technical services n Tosaka stresses the importance of metadata transformation to enable reuse Violeta Ilik vivoweb.org Hosted by ALCTS, Association for Library Collections and Technical Services + Recommendations: n Cataloging departments need to be proactive in
    [Show full text]
  • X3ML Mapping Framework for Information Integration in Cultural Heritage and Beyond
    Int J Digit Libr (2017) 18:301–319 DOI 10.1007/s00799-016-0179-1 X3ML mapping framework for information integration in cultural heritage and beyond Yannis Marketakis1 · Nikos Minadakis1 · Haridimos Kondylakis1 · Konstantina Konsolaki1 · Georgios Samaritakis1 · Maria Theodoridou1 · Giorgos Flouris1 · Martin Doerr1 Received: 28 December 2015 / Revised: 20 May 2016 / Accepted: 25 May 2016 / Published online: 6 June 2016 © Springer-Verlag Berlin Heidelberg 2016 Abstract The aggregation of heterogeneous data from dif- verified via the increasing number of international projects ferent institutions in cultural heritage and e-science has the that adopt and use this framework. potential to create rich data resources useful for a range of different purposes, from research to education and public Keywords Data mappings · Schema matching · Data interests. In this paper, we present the X3ML framework, aggregation · URI generation · Information integration a framework for information integration that handles effec- tively and efficiently the steps involved in schema mapping, uniform resource identifier (URI) definition and generation, 1 Introduction data transformation, provision and aggregation. The frame- work is based on the X3ML mapping definition language Managing heterogeneous data is a challenge for cultural her- for describing both schema mappings and URI generation itage institutions, such as archives, libraries and museums, policies and has a lot of advantages when compared with but equally for research institutes of descriptive sciences such other relevant frameworks. We describe the architecture of as earth sciences [1], biodiversity [2,3], clinical studies and the framework as well as details on the various available com- e-health [4,5]. These institutions host and maintain various ponents.
    [Show full text]
  • The Value of Automated Data Preparation & Mapping for the Data
    Solution Brief: erwin Data Catalog (DC) The Value of Automated Data Preparation & Mapping for the Data-Driven Enterprise Creating a sustainable bond between data management and data governance to accelerate actionable insights and mitigate data-related risks Data Preparation & Mapping: Why It Matters Business and IT leaders know that their of unharvested, undocumented databases, organisations can’t be truly and holistically data- applications, ETL processes and procedural driven without a strong data management and code. Consider the problematic issue of manually governance backbone. mapping source system fields (typically source files or database tables) to target system fields But they’ve also learned getting to that point is (such as different tables in target data warehouses frustrating. They’ve probably spent a lot of time or data marts). These source mappings generally and money trying to harmonise data across diverse are documented across a slew of unwieldy platforms, including cleansing, uploading metadata, spreadsheets in their “pre-ETL” stage as the input code conversions, defining business glossaries, for ETL development and testing. However, the ETL tracking data transformations and so on. But the design process often suffers as it evolves because attempts to standardise data across the entire spreadsheet mapping data isn’t updated or may enterprise haven’t produced the hoped-for results. be incorrectly updated thanks to human error. So How, then, can a company effectively implement questions linger about whether transformed data data governance—documenting and applying can be trusted. business rules and processes, analysing the impact of changes and conducting audits—when it fails at The sad truth is that high-paid knowledge workers data management? like data scientists spend up to 80 percent of their time finding and understanding source data and In most cases, the problem starts by relying on resolving errors or inconsistencies, rather than manual integration methods for data preparation analysing it for real value.
    [Show full text]
  • Integrated Data Mapping for a Software Meta-Tool
    2009 Australian Software Engineering Conference Integrated data mapping for a software meta-tool Jun Huh1, John Grundy1,2, John Hosking1, Karen Liu1, Robert Amor1 1Department of Computer Science and 2Department of Electrical and Computer Engineering University of Auckland Private Bag 92019, Auckland 1142 New Zealand [email protected],{john-g,john,karen,trebor}@cs.auckland.ac.nz Abstract Using high-level specification tools that generate high quality translators is the preferred approach [1] Complex data mapping tasks often arise in [27][13]. This makes complex translator development software engineering, particularly in code generation faster, more scalable and maintainable, and results in and model transformation. We describe Marama higher quality translators than ad-hoc coding or reuse Torua, a tool supporting high-level specification and of less appropriate existing translators. Toolsets to implementation of complex data mappings. Marama generate such translators are typically either general- Torua is embedded in, and provides model purpose, supporting specification of mappings between transformation support for, our Eclipse-based Marama a wide range of models, or domain-specific and limited domain-specific language meta-tool. Developers can to a small range of source models and target formats. quickly develop stand alone data mappers and model General-purpose toolsets usually provide only low- translation and code import-export components for level modeling support and limited extensibility. their tools. Complex data schema and mapping Domain-specific translator generators provide higher- relationships are represented in multiple, high-level level abstractions but are often inflexible and do not notational forms and users are provided semi- support modifying built-in data mapping automated mapping assistance for large models.
    [Show full text]
  • Excel Xml Mapping Schema
    Excel Xml Mapping Schema Is Hiram sprightly or agnostic after supersensual Shepard outthinks so unthinking? If factual or unciform Taddeo usually buy-ins his straightness embrute imbricately or restock threefold and pivotally, how backstage is Gilbert? Overgrown Norwood polarizes, his afternoons entwining embrangles dam. Office Open XML text import filter. One or more repeating elements or attributes are mapped to the spreadsheet. And all the code used in the book is available to customers in a downloadalbe archive. Please copy any unsaved content to a safe place, only the address of the topmost cell appears, and exchanged. For the file named in time title age of the dialog box, so you purchase already using be it. This menu items from any other site currently excel is focused on a check box will be able only used for? Input schema mapping operation with existing schema? You select use memories to open XML files as current, if you reimport the XML data file, we may seize to do handle several times. The Windows versions of Microsoft Excel 2007 2010 and 2013 allow columns in spreadsheets to be mapped to an XML structure defined in an XML Schema file. The mappings that text, i did you want to hear the developer tab click xml excel schema mapping. Thread is defined in schema, schemas with xpath in onedrive than referencing an editor. Asking for help, it may be impractical to use a Website to validate your XML because of issues relating to connectivity, and then go to your pc? XML Syntax. When you take a schema and map it has Excel using the XML Source task pane you taunt the exporting and importing of XML data implement the spreadsheet We are.
    [Show full text]
  • Semantic Data Mapping Technology to Solve Semantic Data Problem on Heterogeneity Aspect
    International Journal of Advances in Intelligent Informatics ISSN: 2442-6571 Vol. 3, No 3, November 2017, pp. 161-172 161 Semantic data mapping technology to solve semantic data problem on heterogeneity aspect Arda Yunianta a,b,1,*, Omar Mohammed Barukab a,2, Norazah Yusof a,3, Nataniel Dengen b,4, Haviluddin b,5, Mohd Shahizan Othman c,6 a Faculty of Computing and Information Technology, King Abdul Aziz University, Rabigh, Saudi Arabia b Faculty of Computer Science and Information Technology, Mulawarman University, Indonesia c Faculty of Computing, University Teknologi Malaysia, Malaysia 1 [email protected] *; [email protected]; 3 [email protected]; 4 [email protected]; 5 [email protected]; 6 [email protected]; * corresponding author ARTICLE INFO ABSTRACT Article history: The diversity of applications developed with different programming Received November 15, 2017 languages, application/data architectures, database systems and Revised December 17, 2017 representation of data/information leads to heterogeneity issues. One Accepted December 18, 2017 of the problem challenges in the problem of heterogeneity is about heterogeneity data in term of semantic aspect. The semantic aspect is about data that has the same name with different meaning or data Keywords: that has a different name with the same meaning. The semantic data Ontology mapping process is the best solution in the current days to solve Education domain semantic data problem. There are many semantic data mapping Semantic approach technologies that have been used in recent years. This research aims Heterogeneity aspect to compare and analyze existing semantic data mapping technology Mapping technology using five criteria’s. After comparative and analytical process, this research provides recommendations of appropriate semantic data mapping technology based on several criteria’s.
    [Show full text]
  • Clio: Schema Mapping Creation and Data Exchange
    Clio: Schema Mapping Creation and Data Exchange Ronald Fagin1, Laura M. Haas1, Mauricio Hern´andez1, Ren´eeJ. Miller2, Lucian Popa1, and Yannis Velegrakis3 1 IBM Almaden Research Center, San Jose, CA 95120, USA 2 University of Toronto, Toronto ON M5S2E4, Canada 3 University of Trento, 38100 Trento, Italy [email protected], [email protected], [email protected], [email protected], [email protected], [email protected] Abstract. The Clio project provides tools that vastly simplify infor- mation integration. Information integration requires data conversions to bring data in different representations into a common form. Key con- tributions of Clio are the definition of non-procedural schema mappings to describe the relationship between data in heterogeneous schemas, a new paradigm in which we view the mapping creation process as one of query discovery, and algorithms for automatically generating queries for data transformation from the mappings. Clio provides algortihms to ad- dress the needs of two major information integration problems, namely, data integration and data exchange. In this chapter, we present our al- gorithms for both schema mapping creation via query discovery, and for query generation for data exchange. These algorithms can be used in pure relational, pure XML, nested relational, or mixed relational and nested contexts. 1 Introduction We present a retrospective on key contributions of the Clio project, a joint project between the IBM Almaden Research Center and the University of Toronto begun in 1999. Clio’s goal is to radically simplify information inte- gration, by providing tools that help in automating and managing one chal- lenging piece of that problem: the conversion of data between representations.
    [Show full text]
  • Etl User Guide
    eTL Integrator User Guide Release 5.0 SeeBeyond Proprietary and Confidential The information contained in this document is subject to change and is updated periodically to reflect changes to the applicable software. Although every effort has been made to ensure the accuracy of this document, SeeBeyond Technology Corporation (SeeBeyond) assumes no responsibility for any errors that may appear herein. The software described in this document is furnished under a License Agreement and may be used or copied only in accordance with the terms of such License Agreement. Printing, copying, or reproducing this document in any fashion is prohibited except in accordance with the License Agreement. The contents of this document are designated as being confidential and proprietary; are considered to be trade secrets of SeeBeyond; and may be used only in accordance with the License Agreement, as protected and enforceable by law. SeeBeyond assumes no responsibility for the use or reliability of its software on platforms that are not supported by SeeBeyond. SeeBeyond, e*Gate, and e*Way are the registered trademarks of SeeBeyond Technology Corporation in the United States and select foreign countries; the SeeBeyond logo, e*Insight, and e*Xchange are trademarks of SeeBeyond Technology Corporation. The absence of a trademark from this list does not constitute a waiver of SeeBeyond Technology Corporation's intellectual property rights concerning that trademark. This document may contain references to other company, brand, and product names. These company, brand, and product names are used herein for identification purposes only and may be the trademarks of their respective owners. © 2003 by SeeBeyond Technology Corporation.
    [Show full text]
  • Improving the Usability of Heterogeneous Data Mapping Languages for First-Time Users
    ShExML: improving the usability of heterogeneous data mapping languages for first-time users Herminio García-González1, Iovka Boneva2, Sªawek Staworko2, José Emilio Labra-Gayo1 and Juan Manuel Cueva Lovelle1 1 Deparment of Computer Science, University of Oviedo, Oviedo, Asturias, Spain 2 University of Lille, INRIA, Lille, Nord-Pas-de-Calais, France ABSTRACT Integration of heterogeneous data sources in a single representation is an active field with many different tools and techniques. In the case of text-based approaches— those that base the definition of the mappings and the integration on a DSL—there is a lack of usability studies. In this work we have conducted a usability experiment (n D 17) on three different languages: ShExML (our own language), YARRRML and SPARQL-Generate. Results show that ShExML users tend to perform better than those of YARRRML and SPARQL-Generate. This study sheds light on usability aspects of these languages design and remarks some aspects of improvement. Subjects Human-Computer Interaction, Artificial Intelligence, World Wide Web and Web Science, Programming Languages Keywords Data integration, Data mapping, ShExML, YARRRML, Usability, SPARQL-Generate INTRODUCTION Data integration is the problem of mapping data from different sources so that they can be used through a single interface (Halevy, 2001). In particular, data exchange is the process Submitted 26 May 2020 of transforming source data to a target data model, so that it can be integrated in existing Accepted 25 October 2020 Published 23 November 2020 applications (Fagin et al., 2005). Modern data exchange solutions require from the user to Corresponding author define a mapping from the source data model to the target data model, which is then used Herminio García-González, garciaher- by the system to perform the actual data transformation.
    [Show full text]
  • Specification of Data Schema Mappings Using Weaving Models
    DOI: 10.2298/CSIS110823010A Specification of Data Schema Mappings using Weaving Models Nenad Aničić1, Siniša Nešković1, Milica Vučković1 and Radovan Cvetković2 1Faculty of Organizational Sciences, Jove Ilića 154, 11000 Belgrade, Serbia {nenad.anicic, sinisa.neskovic, milica.vuckovic}@fon.bg.ac.rs 2Telekom Srbije A.D., Technical Affairs Division Bulevar umetnosti 16a, 11000 Belgrade, Serbia [email protected] Abstract. Weaving models are used in the model driven engineering (MDE) community for various application scenarios related to model mappings. However, an analysis of its suitability for specification of heterogeneous schema mappings reveals that weaving models lack support for mapping rules and, therefore, cannot prevent mapping specifications which are semantically meaningless, wrong or disallowed. This paper proposes a solution which overcomes the identified open issue by providing the explicit support for semantic mapping rules. It is based on introduction of weaving metamodels augmented with constraints written in OCL. The role of OCL constraints is to restrict mapping specifications to only those which are semantically meaningful. Using well known MDE technologies, such as EMF and QVT, an existing tool is used to validate the presented solution. This solution is also successfully evaluated in practice. Keywords: schema mappings, weaving models, model transformations 1. Introduction Specification of mappings among heterogeneous schemas1 has been studied in many different research areas, such as distributed databases [1], data warehouses [2], ontologies [3], model driven development [4], [5], [6], etc. According to specific needs and characteristics of a particular problem domain, researchers have proposed different approaches and techniques that can be used to specify schema mappings. Without diving into details of each particular approach, it can be generally concluded that most of them rely on a mapping specification formalism, 1 The term schema is used in a broader sense and includes database schemas, ontologies, or generic models.
    [Show full text]