Modeling and Management Big Data in Databases—A Systematic Literature Review

Total Page:16

File Type:pdf, Size:1020Kb

Modeling and Management Big Data in Databases—A Systematic Literature Review sustainability Review Modeling and Management Big Data in Databases—A Systematic Literature Review Diana Martinez-Mosquera 1,* , Rosa Navarrete 1 and Sergio Lujan-Mora 2 1 Department of Informatics and Computer Science, Escuela Politécnica Nacional, 170525 Quito, Ecuador; [email protected] 2 Department of Software and Computing Systems, University of Alicante, 03690 Alicante, Spain; [email protected] * Correspondence: [email protected]; Tel.: +593-02-2976300 Received: 25 November 2019; Accepted: 9 January 2020; Published: 15 January 2020 Abstract: The work presented in this paper is motivated by the acknowledgement that a complete and updated systematic literature review (SLR) that consolidates all the research efforts for Big Data modeling and management is missing. This study answers three research questions. The first question is how the number of published papers about Big Data modeling and management has evolved over time. The second question is whether the research is focused on semi-structured and/or unstructured data and what techniques are applied. Finally, the third question determines what trends and gaps exist according to three key concepts: the data source, the modeling and the database. As result, 36 studies, collected from the most important scientific digital libraries and covering the period between 2010 and 2019, were deemed relevant. Moreover, we present a complete bibliometric analysis in order to provide detailed information about the authors and the publication data in a single document. This SLR reveal very interesting facts. For instance, Entity Relationship and document-oriented are the most researched models at the conceptual and logical abstraction level respectively and MongoDB is the most frequent implementation at the physical. Furthermore, 2.78% studies have proposed approaches oriented to hybrid databases with a real case for structured, semi-structured and unstructured data. Keywords: big data; management; modeling; literature review 1. Introduction The Big Data modeling term became widespread in 2011, as is visible in Figure1. This figure shows searches in Google Trends related to Big Data modeling, which are intensified from 2011 onwards. Searches before 2004 are not presented, since Google Trends does not store earlier data. In recent years, researchers have consolidated their efforts to study new paradigms to deal with Big Data. Thus, novel Big Data modeling and management in databases approaches have emerged, in line with the new requirements. In consequence, new techniques in the database context have evolved towards Not Only SQL (NoSQL). The work presented in this paper is motivated by the acknowledgement that a complete systematic literature review (SLR) that consolidates all the research efforts for Big Data modeling and management in databases is missing. An SLR is the best way to collect, summarize and evaluate all scientific evidence about a topic [1]. It allows for the description of research areas shown the greatest and least interest by researchers. Considering the exposed issues, the SLR conducted in this work can contribute to solving this lack by collecting and analyzing details about the research published from 2010 to 2019. As a basis for our SLR, we adhered to the guidelines proposed by Kitchenham [1]. Moreover, this paper presents a complete bibliometric analysis and summarizes existing evidence about research Sustainability 2020, 12, 634; doi:10.3390/su12020634 www.mdpi.com/journal/sustainability Sustainability 2020, 12, 634 2 of 41 Sustainability 2019, 11, x FOR PEER REVIEW 2 of 44 on Big Data modeling and management in databases. With the information attained from the analysis performed,from the analysis we will performed, identify trends we wi andll identify gaps in thetrends published and gaps research in the topublished provide aresearch background to provide for new a research.background As result,for new 1376 research. papers As were result, obtained 1376 frompapers scientific were obtained libraries from and 36 scientific studies werelibraries selected and 36 as relevant.studies were All theselected research as relevant. efforts were All mappedthe research in order efforts to respondwere mapped the three in order research to respond questions the defined three inresearch this research. questions Our defined main goal in this is toresearch. consolidate Our main the main goal works is to consolidate to provide the an awarenessmain works of to the provide trends andan awareness the gaps relatedof the trends to Big Dataand the modeling. gaps related to Big Data modeling. The remainder ofof this this paper paper is is organized organized as as follows. follows. First First is theis the Introduction Introduction section, section, containing containing the meaningthe meaning of the of dithefferent different terms terms discussed discussed in this in study. this study. Second, Second, a Method a Method section section presents presents the process the usedprocess to performused to theperform planning, the planning, conducting conducting and reporting and of reporting the SLR. Theof the planning SLR. The phase planning describes phase the identificationdescribes the identification of the need for of a the SLR need study for and a SL theR study development and the ofdevelopment a review protocol, of a review objectives protocol, and justification,objectives and research justification, questions research and strategy. questions The conductingand strategy. phase The presents conducti theng inclusion phase andpresents selection the criteriainclusion for and the selection final corpus criteria of selected for the studies.final corpus of selected studies. Third, thethe Results Results section section answers answers the researchthe research questions questions in three in subsections. three subsections. The first subsection, The first thesubsection, Bibliometric the Bibliometric Analysis, comprises Analysis, comprises the reporting the reporting stage, including stage, including authors’ authors’ information, information, such as asuchffiliation as affiliation and country and andcountry relevant and data relevant about data their about works; their for instance,works; for publication instance, information,publication numberinformation, of citations, number funding of citations, source, funding year, digital source, library, year, impact digital factor, library, ranking. impact The factor, second ranking. subsection, The thesecond Systematic subsection, Literature the Systematic Review, presents Literature a mapping Review, ofpresents the selected a mapping studies of according the selected to three studies key conceptsaccording in to athree concept key matrixconcepts regarding in a concept to the matr datasetix regarding source, to modeling the dataset and source, database. modeling The third and subsection,database. The the third Discussion, subsection, highlights the Discussion, relevant findingshighlights in relevant the SLR findings study in in order the SLR to identify study in existing order trendsto identify and existing gaps. Finally, trends conclusions and gaps. Finally and future, conclusions works are and presented. future works are presented. Figure 1. TrendTrend of use of term Big Data [[2].2]. 1.1. Big Data Concepts In this part, we describe the main concepts related to Big Data, in order to provide a general overview for the reader and a backgroundbackground ofof thethe termsterms discusseddiscussed later.later. 1.1.1. A Brief History of Big Data 1.1.1. A Brief History of Big Data The production and processing of large volumes of data began to be of interest to researchers many The production and processing of large volumes of data began to be of interest to researchers years ago. By 1944, estimations for the size of libraries, which increased rapidly every year, were made many years ago. By 1944, estimations for the size of libraries, which increased rapidly every year, in American universities [3]. In 1997, at the Institute of Electrical and Electronics Engineers (IEEE) were made in American universities [3]. In 1997, at the Institute of Electrical and Electronics Engineers Conference on Visualization, the term “Big Data” was used for the first time during the presentation of (IEEE) Conference on Visualization, the term “Big Data” was used for the first time during the a study about large datasets’ visualization [4]. presentation of a study about large datasets’ visualization [4]. Big Data is the buzzword of recent years, that is, a fashionable expression in information systems. Big Data is the buzzword of recent years, that is, a fashionable expression in information The general population relates the term Big Data to its literal meaning of large volumes of data. systems. The general population relates the term Big Data to its literal meaning of large volumes of However, Big Data is a generic term used to refer to large and complex datasets that arise from the data. However, Big Data is a generic term used to refer to large and complex datasets that arise from combination of famous Big Data V’s that characterize it [5]. the combination of famous Big Data V’s that characterize it [5]. 1.1.2. Big Data Characterization 1.1.2. Big Data Characterization As mentioned before, Big Data does not refer only to high volumes of data to be processed. At the As mentioned before, Big Data does not refer only to high volumes of data to be processed. At beginning of the Big Data studies, their volume, velocity and variety were considered as fundamental the beginning of
Recommended publications
  • Data Warehouse: an Integrated Decision Support Database Whose Content Is Derived from the Various Operational Databases
    1 www.onlineeducation.bharatsevaksamaj.net www.bssskillmission.in DATABASE MANAGEMENT Topic Objective: At the end of this topic student will be able to: Understand the Contrasting basic concepts Understand the Database Server and Database Specified Understand the USER Clause Definition/Overview: Data: Stored representations of objects and events that have meaning and importance in the users environment. Information: Data that have been processed in such a way that they can increase the knowledge of the person who uses it. Metadata: Data that describes the properties or characteristics of end-user data and the context of that data. Database application: An application program (or set of related programs) that is used to perform a series of database activities (create, read, update, and delete) on behalf of database users. WWW.BSSVE.IN Data warehouse: An integrated decision support database whose content is derived from the various operational databases. Constraint: A rule that cannot be violated by database users. Database: An organized collection of logically related data. Entity: A person, place, object, event, or concept in the user environment about which the organization wishes to maintain data. Database management system: A software system that is used to create, maintain, and provide controlled access to user databases. www.bsscommunitycollege.in www.bssnewgeneration.in www.bsslifeskillscollege.in 2 www.onlineeducation.bharatsevaksamaj.net www.bssskillmission.in Data dependence; data independence: With data dependence, data descriptions are included with the application programs that use the data, while with data independence the data descriptions are separated from the application programs. Data warehouse; data mining: A data warehouse is an integrated decision support database, while data mining (described in the topic introduction) is the process of extracting useful information from databases.
    [Show full text]
  • How Big Data Enables Economic Harm to Consumers, Especially to Low-Income and Other Vulnerable Sectors of the Population
    How Big Data Enables Economic Harm to Consumers, Especially to Low-Income and Other Vulnerable Sectors of the Population The author of these comments, Nathan Newman, has been writing for twenty years about the impact of technology on society, including his 2002 book Net Loss: Internet Profits, Private Profits and the Costs to Community, based on doctoral research on rising regional economic inequality in Silicon Valley and the nation and the role of Internet policy in shaping economic opportunity. He has been writing extensively about big data as a research fellow at the New York University Information Law Institute for the last two years, including authoring two law reviews this year on the subject. These comments are adapted with some additional research from one of those law reviews, "The Costs of Lost Privacy: Consumer Harm and Rising Economic Inequality in the Age of Google” published in the William Mitchell Law Review. He has a J.D. from Yale Law School and a Ph.D. from UC-Berkeley’s Sociology Department; Executive Summary: ϋBͤ͢ Dͯ͜͜ό ͫͧͯͪͭͨͮ͜͡ ͮͰͣ͞ ͮ͜ Gͪͪͧ͢͠ ͩ͜͟ Fͪͪͦ͜͞͠͝ ͭ͜͠ ͪͨͤͩ͢͝͠͞ ͪͨͤͩͩͯ͟͜ institutions organizing information not just about the world but about consumers themselves, thereby reshaping a range of markets based on empowering a narrow set of corporate advertisers and others to prey on consumers based on behavioral profiling. While big data can benefit consumers in certain instances, there are a range of new consumer harms to users from its unregulated use by increasingly centralized data platforms. A. These harms start with the individual surveillance of users by employers, financial institutions, the government and other players that these platforms allow, but also extend to ͪͭͨͮ͡ ͪ͡ Ͳͣͯ͜ ͣͮ͜ ͩ͝͠͠ ͧͧ͜͟͞͠ ϋͧͪͭͤͯͣͨͤ͜͢͞ ͫͭͪͤͧͤͩ͢͡ό ͯͣͯ͜ ͧͧͪ͜Ͳ ͪͭͫͪͭͯ͜͞͠ ͤͩͮͯͤͯͰͯͤͪͩͮ to discriminate and exploit consumers as categorical groups.
    [Show full text]
  • Powerdesigner 16.6 Data Modeling
    SAP® PowerDesigner® Document Version: 16.6 – 2016-02-22 Data Modeling Content 1 Building Data Models ...........................................................8 1.1 Getting Started with Data Modeling...................................................8 Conceptual Data Models........................................................8 Logical Data Models...........................................................9 Physical Data Models..........................................................9 Creating a Data Model.........................................................10 Customizing your Modeling Environment........................................... 15 1.2 Conceptual and Logical Diagrams...................................................26 Supported CDM/LDM Notations.................................................27 Conceptual Diagrams.........................................................31 Logical Diagrams............................................................43 Data Items (CDM)............................................................47 Entities (CDM/LDM)..........................................................49 Attributes (CDM/LDM)........................................................55 Identifiers (CDM/LDM)........................................................58 Relationships (CDM/LDM)..................................................... 59 Associations and Association Links (CDM)..........................................70 Inheritances (CDM/LDM)......................................................77 1.3 Physical Diagrams..............................................................82
    [Show full text]
  • MDA Process to Extract the Data Model from Document-Oriented Nosql Database
    MDA Process to Extract the Data Model from Document-oriented NoSQL Database Amal Ait Brahim, Rabah Tighilt Ferhat and Gilles Zurfluh Toulouse Institute of Computer Science Research (IRIT), Toulouse Capitole University, Toulouse, France Keywords: Big Data, NoSQL, Model Extraction, Schema Less, MDA, QVT. Abstract: In recent years, the need to use NoSQL systems to store and exploit big data has been steadily increasing. Most of these systems are characterized by the property "schema less" which means absence of the data model when creating a database. This property brings an undeniable flexibility by allowing the evolution of the model during the exploitation of the base. However, query expression requires a precise knowledge of the data model. In this article, we propose a process to automatically extract the physical model from a document- oriented NoSQL database. To do this, we use the Model Driven Architecture (MDA) that provides a formal framework for automatic model transformation. From a NoSQL database, we propose formal transformation rules with QVT to generate the physical model. An experimentation of the extraction process was performed on the case of a medical application. 1 INTRODUCTION however that it is absent in some systems such as Cassandra and Riak TS. The "schema less" property Recently, there has been an explosion of data offers undeniable flexibility by allowing the model to generated and accumulated by more and more evolve easily. For example, the addition of new numerous and diversified computing devices. attributes in an existing line is done without Databases thus constituted are designated by the modifying the other lines of the same type previously expression "Big Data" and are characterized by the stored; something that is not possible with relational so-called "3V" rule (Chen, 2014).
    [Show full text]
  • Issues, Challenges, and Solutions: Big Data Mining
    ISSUES , CHALLENGES , AND SOLUTIONS : BIG DATA MINING Jaseena K.U. 1 and Julie M. David 2 1,2 Department of Computer Applications, M.E.S College, Marampally, Aluva, Cochin, India [email protected],[email protected] ABSTRACT Data has become an indispensable part of every economy, industry, organization, business function and individual. Big Data is a term used to identify the datasets that whose size is beyond the ability of typical database software tools to store, manage and analyze. The Big Data introduce unique computational and statistical challenges, including scalability and storage bottleneck, noise accumulation, spurious correlation and measurement errors. These challenges are distinguished and require new computational and statistical paradigm. This paper presents the literature review about the Big data Mining and the issues and challenges with emphasis on the distinguished features of Big Data. It also discusses some methods to deal with big data. KEYWORDS Big data mining, Security, Hadoop, MapReduce 1. INTRODUCTION Data is the collection of values and variables related in some sense and differing in some other sense. In recent years the sizes of databases have increased rapidly. This has lead to a growing interest in the development of tools capable in the automatic extraction of knowledge from data [1]. Data are collected and analyzed to create information suitable for making decisions. Hence data provide a rich resource for knowledge discovery and decision support. A database is an organized collection of data so that it can easily be accessed, managed, and updated. Data mining is the process discovering interesting knowledge such as associations, patterns, changes, anomalies and significant structures from large amounts of data stored in databases, data warehouses or other information repositories.
    [Show full text]
  • Using BEAM Software to Simulate the Introduction of On-Demand, Automated, and Electric Shuttles for Last Mile Connectivity in Santa Clara County
    San Jose State University SJSU ScholarWorks Mineta Transportation Institute Publications 1-2021 Using BEAM Software to Simulate the Introduction of On-Demand, Automated, and Electric Shuttles for Last Mile Connectivity in Santa Clara County Gary Hsueh Mineta Transportation Institute David Czerwinski San Jose State University Cristian Poliziani Mineta Transportation Institute Terris Becker San Jose State University Alexandre Hughes San Jose State University See next page for additional authors Follow this and additional works at: https://scholarworks.sjsu.edu/mti_publications Part of the Sustainability Commons, and the Transportation Commons Recommended Citation Gary Hsueh, David Czerwinski, Cristian Poliziani, Terris Becker, Alexandre Hughes, Peter Chen, and Melissa Benn. "Using BEAM Software to Simulate the Introduction of On-Demand, Automated, and Electric Shuttles for Last Mile Connectivity in Santa Clara County" Mineta Transportation Institute Publications (2021). https://doi.org/10.31979/mti.2021.1822 This Report is brought to you for free and open access by SJSU ScholarWorks. It has been accepted for inclusion in Mineta Transportation Institute Publications by an authorized administrator of SJSU ScholarWorks. For more information, please contact [email protected]. Authors Gary Hsueh, David Czerwinski, Cristian Poliziani, Terris Becker, Alexandre Hughes, Peter Chen, and Melissa Benn This report is available at SJSU ScholarWorks: https://scholarworks.sjsu.edu/mti_publications/343 Project 1822 January 2021 Using BEAM Software to
    [Show full text]
  • Integration Definition for Function Modeling (IDEF0)
    NIST U.S. DEPARTMENT OF COMMERCE PUBLICATIONS £ Technology Administration National Institute of Standards and Technology FIPS PUB 183 FEDERAL INFORMATION PROCESSING STANDARDS PUBLICATION INTEGRATION DEFINITION FOR FUNCTION MODELING (IDEFO) » Category: Software Standard SUBCATEGORY: MODELING TECHNIQUES 1993 December 21 183 PUB FIPS JK- 45C .AS A3 //I S3 IS 93 FIPS PUB 183 FEDERAL INFORMATION PROCESSING STANDARDS PUBLICATION INTEGRATION DEFINITION FOR FUNCTION MODELING (IDEFO) Category: Software Standard Subcategory: Modeling Techniques Computer Systems Laboratory National Institute of Standards and Technology Gaithersburg, MD 20899 Issued December 21, 1993 U.S. Department of Commerce Ronald H. Brown, Secretary Technology Administration Mary L. Good, Under Secretary for Technology National Institute of Standards and Technology Arati Prabhakar, Director Foreword The Federal Information Processing Standards Publication Series of the National Institute of Standards and Technology (NIST) is the official publication relating to standards and guidelines adopted and promulgated under the provisions of Section 111 (d) of the Federal Property and Administrative Services Act of 1949 as amended by the Computer Security Act of 1987, Public Law 100-235. These mandates have given the Secretary of Commerce and NIST important responsibilities for improving the utilization and management of computer and related telecommunications systems in the Federal Government. The NIST, through its Computer Systems Laboratory, provides leadership, technical guidance,
    [Show full text]
  • Drawing-A-Database-Schema.Pdf
    Drawing A Database Schema Padraig roll-out her osteotome pluckily, trillion and unacquainted. Astronomic Dominic haemorrhage operosely. Dilative Parrnell jury-rigging: he bucketing his sympatholytics tonishly and litho. Publish your schema. And database user schema of databases in berlin for your drawing created in a diagram is an er diagram? And you know some they say, before what already know. You can generate the DDL and modify their hand for SQLite, although to it ugly. How can should improve? This can work online, a record is crucial to reduce faults in. The mouse pointer should trace to an icon with three squares. Visual Database Creation with MySQL Workbench Code. In database but a schema pronounced skee-muh or skee-mah is the organisation and structure of a syringe Both schemas and. Further more complex application performance, concept was that will inform your databases to draw more control versions. Typically goes in a schema from any sql for these terms of maintenance of the need to do you can. Or database schemas you draw data models commonly used to select all databases by drawing page helpful is in a good as methods? It is far to bath to target what suits you best. Gallery of training courses. Schema for database schema for. Help and Training on mature site? You can jump of ER diagrams as a simplified form let the class diagram and carpet may be easier for create database design team members to. This token will be enrolled in quickly create drawings by enabled the left side of the process without realising it? Understanding a Schema in Psychology Verywell Mind.
    [Show full text]
  • A Pattern-Based Approach to Ontology-Supported Semantic Data Integration
    A pattern-based approach to ontology-supported semantic data integration Andreas A. Jordan Wolfgang Mayer Georg Grossmann Markus Stumptner Advanced Computing Research Centre University of South Australia, Mawson Lakes Boulevard, Mawson Lakes, South Australia 5095, Email: [email protected] Email: (wolfgang.mayer|georg.grossmann)@unisa.edu.au, [email protected] Abstract RSM2, ISO 159263 and Gellish4 standards are all used to represent information about engineering assets and Interoperability between information systems is associated maintenance information. However, there rapidly becoming an increasingly difficult issue, in exists considerable variation in the scope and level of particular where data is to be exchanged with exter- detail present in each standard, and where the infor- nal entities. Although standard naming conventions mation content covered by different standards over- and forms of representation have been established, laps, their vocabulary and representation used vary significant discrepancies remain between different im- considerably. Furthermore, standards like ISO 15926 plementations of the same standards and between dif- are large yet offer little guidance on how to represent ferent standards within the same domain. Further- particular asset information in its generic data struc- more, current approaches for bridging such hetero- tures. Heterogeneities arising from continued evolu- geneities are incomplete or do not scale well to real- tion of standards have further exacerbated the prob- world problems. lem. As a consequence, tool vendors formally support We present a pragmatic approach to interoperabil- the standard, yet their implementations do not inter- ity, where ontologies are leveraged to annotate proto- operate due to different use of the same standard or typical information fragments in order to facilitate varying assumptions about the required and optional automated transformation between fragments in dif- information held in each system.
    [Show full text]
  • Oracle Big Data SQL Release 4.1
    ORACLE DATA SHEET Oracle Big Data SQL Release 4.1 The unprecedented explosion in data that can be made useful to enterprises – from the Internet of Things, to the social streams of global customer bases – has created a tremendous opportunity for businesses. However, with the enormous possibilities of Big Data, there can also be enormous complexity. Integrating Big Data systems to leverage these vast new data resources with existing information estates can be challenging. Valuable data may be stored in a system separate from where the majority of business-critical operations take place. Moreover, accessing this data may require significant investment in re-developing code for analysis and reporting - delaying access to data as well as reducing the ultimate value of the data to the business. Oracle Big Data SQL enables organizations to immediately analyze data across Apache Hadoop, Apache Kafka, NoSQL, object stores and Oracle Database leveraging their existing SQL skills, security policies and applications with extreme performance. From simplifying data science efforts to unlocking data lakes, Big Data SQL makes the benefits of Big Data available to the largest group of end users possible. KEY FEATURES Rich SQL Processing on All Data • Seamlessly query data across Oracle Oracle Big Data SQL is a data virtualization innovation from Oracle. It is a new Database, Hadoop, object stores, architecture and solution for SQL and other data APIs (such as REST and Node.js) on Kafka and NoSQL sources disparate data sets, seamlessly integrating data in Apache Hadoop, Apache Kafka, • Runs all Oracle SQL queries without modification – preserving application object stores and a number of NoSQL databases with data stored in Oracle Database.
    [Show full text]
  • Big Data, Data Mining, and Machine Learning: Value Creation for Business Leaders and Practitioners
    Additional praise for Big Data, Data Mining, and Machine Learning: Value Creation for Business Leaders and Practitioners “Jared’s book is a great introduction to the area of High Powered Analytics. It will be useful for those who have experience in predictive analytics but who need to become more versed in how technology is changing the capabilities of existing methods and creating new pos- sibilities. It will also be helpful for business executives and IT profes- sionals who’ll need to make the case for building the environments for, and reaping the benefi ts of, the next generation of advanced analytics.” —Jonathan Levine, Senior Director, Consumer Insight Analysis at Marriott International “The ideas that Jared describes are the same ideas that being used by our Kaggle contest winners. This book is a great overview for those who want to learn more and gain a complete understanding of the many facets of data mining, knowledge discovery and extracting value from data.” —Anthony Goldbloom Founder and CEO of Kaggle “The concepts that Jared presents in this book are extremely valuable for the students that I teach and will help them to more fully under- stand the power that can be unlocked when an organization begins to take advantage of its data. The examples and case studies are particu- larly useful for helping students to get a vision for what is possible. Jared’s passion for analytics comes through in his writing, and he has done a great job of making complicated ideas approachable to multiple audiences.” —Tonya Etchison Balan, Ph.D., Professor of Practice, Statistics, Poole College of Management, North Carolina State University Big Data, Data Mining, and Machine Learning Wiley & SAS Business Series The Wiley & SAS Business Series presents books that help senior-level managers with their critical management decisions.
    [Show full text]
  • Metamodeling and Method Engineering with Conceptbase”
    This is a pre-print of the book chapter M. Jeusfeld: “Metamodeling and method engineering with ConceptBase” . In Jeusfeld, M.A., Jarke, M., Mylopoulos, J. (eds): Metamodeling for Method Engineering, pp. 89-168. The MIT Press., 2009; the original book is available from MIT Press http://mitpress.mit.edu/node/192290 This pre-print may only be used for scholar, non-commercial purposes. Most of the sources for the examples in this chapter are available via http://merkur.informatik.rwth-aachen.de/pub/bscw.cgi/3782591 for download. They require ConceptBase 7.0 or later available from http://conceptbase.cc. Metamodeling and Method Engineering with ConceptBase Manfred Jeusfeld Abstract. This chapter provides a practical guide on how to use the meta data repository ConceptBase to design information modeling methods by using meta- modeling. After motivating the abstraction principles behind meta-modeling, the language Telos as realized in ConceptBase is presented. First, a standard factual representation of statements at any IRDS abstraction level is defined. Then, the foundation of Telos as a logical theory is elaborated yielding simple fixpoint semantics. The principles for object naming, instantiation, attribution, and specialization are reflected by roughly 30 logical axioms. After the language axiomatization, user-defined rules, constraints and queries are introduced. The first part is concluded by a description of active rules that allows the specification of reactions of ConceptBase to external events. The second part applies the language features of the first part to a full-fledged information modeling method: The Yourdan method for Modern Structured Analysis. The notations of the Yourdan method are designed along the IRDS framework.
    [Show full text]