Abstract Real-Time Analytics on Large Dynamic Graphs

Total Page:16

File Type:pdf, Size:1020Kb

Abstract Real-Time Analytics on Large Dynamic Graphs ABSTRACT Title of dissertation: REAL-TIME ANALYTICS ON LARGE DYNAMIC GRAPHS Jayanta Mondal, Doctor of Philosophy, 2015 Dissertation directed by: Professor Amol Deshpande Department of Computer Science In today’s fast-paced and interconnected digital world, the data generated by an increasing number of applications is being modeled as dynamic graphs. The graph struc- ture encodes relationships among data items, while the structural changes to the graphs as well as the continuous stream of information produced by the entities in these graphs make them dynamic in nature. Examples include social networks where users post status updates, images, videos, etc.; phone call networks where nodes may send text messages or place phone calls; road traffic networks where the traffic behavior of the road segments changes constantly, and so on. There is a tremendous value in storing, managing, and analyzing such dynamic graphs and deriving meaningful insights in real-time. However, a majority of the work in graph analytics assumes a static setting, and there is a lack of systematic study of the various dynamic scenarios, the complexity they impose on the analysis tasks, and the challenges in building efficient systems that can support such tasks at a large scale. In this dissertation, I design a unified streaming graph data management framework, and develop prototype systems to support increasingly complex tasks on dynamic graphs. In the first part, I focus on the management and querying of distributed graph data. I de- velop a hybrid replication policy that monitors the read-write frequencies of the nodes to decide dynamically what data to replicate, and whether to do eager or lazy replication in order to minimize network communication and support low-latency querying. In the sec- ond part, I study parallel execution of continuous neighborhood-driven aggregates, where each node aggregates the information generated in its neighborhoods. I build my sys- tem around the notion of an aggregation overlay graph, a pre-compiled data structure that enables sharing of partial aggregates across different queries, and also allows partial pre- computation of the aggregates to minimize the query latencies and increase throughput. Finally, I extend the framework to support continuous detection and analysis of activity- based subgraphs, where subgraphs could be specified using both graph structure as well as activity conditions on the nodes. The query specification tasks in my system are ex- pressed using a set of active structural primitives, which allows the query evaluator to use a set of novel optimization techniques, thereby achieving high throughput. Overall, in this dissertation, I define and investigate a set of novel tasks on dynamic graphs, design scalable optimization techniques, build prototype systems, and show the effectiveness of the proposed techniques through extensive evaluation using large-scale real and synthetic datasets. REAL-TIME ANALYTICS ON LARGE DYNAMIC GRAPHS by Jayanta Mondal Dissertation submitted to the Faculty of the Graduate School of the University of Maryland, College Park in partial fulfillment of the requirements for the degree of Doctor of Philosophy 2015 Advisory Committee: Professor Amol Deshpande, Chair/Advisor Professor Louiqa Raschid, Dean’s Representative Professor V. S. Subrahmanian Professor Hector´ Corrada Bravo Professor Tudor Dumitras, c Copyright by Jayanta Mondal 2015 Acknowledgments I feel greatly humbled in completing this dissertation. I owe my gratitude to all the people who have made this possible through their constant help, guidance, support, and encouragement. Foremost, I want to convey my heartfelt appreciation to my advisor, Dr. Amol Deshpande. It is only due to his patience and support, I have successfully achieved this landmark. He gave me the freedom to explore on my own, and at the same time the guid- ance to stay on the right track. His advice on how to approach problems in a structured fashion and express ideas with clarity, has helped me mature as a researcher. I feel very fortunate to have him as my advisor. Additionally, I would like to acknowledge my committee members V.S. Subrah- manian, Hector Corrada Bravo, Tudor Dumitras, Louiqa Raschid, for their constructive feedback and critiques. Special thanks to Tudor for the wonderful time I had collaborat- ing with him. His guidance has helped me garner a lot of knowledge outside my core area of expertise. I would also like to thank my mentor Dr. Sudipto Das at Microsoft for his guidance during my internships. Working with him has helped me broaden my vision about database research and gain valuable insights about doing research in industry. I would like to take this opportunity to thank the computer science department staff, Jenny Story, Fatima Bangura, Brenda Chick, Sharron McElroy, Jodie Gray, Adelaide Findlay for making my life much simpler by taking care of the dreaded administrative matters. ii Life is nothing without friends. I will cherish the time spent with my grad school friends Rajan, Udayan, Hui, Ashwin, Theodoros, Abdul, Souvik, Amit, Rajesh, Varun, Rajibul, Anirban, Shrin. All the technical and non-technical conversations we had, the meals we shared, the trips to Board and Brew will always be in my memory. This acknowledgement would not be complete without a shout-out to my long-time friends Suman, Sahani, Pravuda, Anushri, Ritwik, Arnab, Kaushani, Moumita, Nagda, Rajarshi, Dibyajyoti, Shaoni. Thank you guys for enriching my life with all the love, memories, conversations, food, and adventures. Special thanks to my girlfriend Jineta for being a great friend, critic, and guide over the years. Above all, I want to thank my parents Sukla and Jaychand and my brother Joydip, for their selfless love and support. Everything I have achieved in my life, I owe every bit to them. iii Table of Contents List of Tables ix List of Figures x 1 Introduction 1 1.1 Data Components of Dynamic Graphs . .6 1.1.1 Network Component . .7 1.1.2 Stream Component . .8 1.1.3 Dependancy between Network and Stream Components . 10 1.2 Task Categories . 11 1.2.1 Large Number of Independent Local Queries . 11 1.2.2 Global Analytics over Content-dynamic Graphs . 12 1.2.3 Event/pattern Detection on Dynamic Graphs . 12 1.2.4 Managing a Large Collection of Small Graphs. 13 1.3 Task/Query Model . 13 1.3.1 Query Issuer . 15 1.3.2 Query Scope . 16 1.3.2.1 Network Traversal Scope . 16 1.3.2.2 Temporal Scope . 18 iv 1.3.3 Answering Model . 20 1.4 System Architecture and Execution Model . 20 1.5 Overview of the Dissertation Research . 26 1.5.1 Distributed Graph Database with Adaptive Replication . 27 1.5.2 Ego-centric Aggregation Framework for Large Dynamic Graphs . 30 1.5.3 A System for Continuous Analytics of Activity-based Subgraphs . 33 1.6 Organization of the Dissertation . 35 2 Related Work 36 2.1 Systems and Techniques for Managing Large Graphs . 36 2.2 Data Stream Management . 41 2.3 Graph Query Languages . 43 2.4 Subgraph Pattern Matching on Dynamic Graphs . 45 3 Distributed Graph Data Management with Adaptive Replication 48 3.1 Overview . 48 3.1.1 Data and Query Model . 49 3.1.2 Architecture . 50 3.1.3 Trade-offs and Requirements . 54 3.2 Replication Manager . 57 3.2.1 Overview . 57 3.2.2 Monitoring Access Patterns . 60 3.2.3 Clustering . 62 3.3 Making Replication Decisions . 63 3.3.1 Problem Definition . 64 3.3.2 Analysis . 65 3.3.3 Proposed Algorithm . 69 3.4 Evaluation . 73 v 3.4.1 Dataset . 74 3.4.2 Experimental Setup . 75 3.4.3 Evaluation Metric . 76 3.4.4 Results . 77 4 Ego-centric Aggregation on Large Dynamic Graphs 85 4.1 Overview . 86 4.1.1 Data and Query Model . 86 4.1.2 Proposed Aggregation Framework . 89 4.1.2.1 Aggregation Overlay Graph . 90 4.1.2.2 Execution Model . 92 4.1.2.3 User-defined Aggregate API . 94 4.2 Constructing The Overlay . 95 4.2.1 Preliminaries . 96 4.2.2 Overlay Construction Algorithms . 98 4.2.3 Handling Dynamic Changes . 107 4.3 Making Dataflow Decisions . 109 4.4 Evaluation . 121 4.4.1 Experimental Setup . 122 4.4.2 Overlay Construction . 124 4.4.3 Dataflow Decisions . 127 4.4.4 Throughput Comparison . 129 5 Activity-based Subgraph Pattern Matching Queries 134 5.1 CASQD Overview . 134 5.1.1 Background . 135 5.1.2 Data Model . 137 5.1.3 Query Model . 140 vi 5.1.3.1 Specifying Active Subgraph Pattern . 140 5.1.3.2 Analytical Task . 143 5.1.4 Answer Reporting Model . 144 5.1.5 Computation Model . 144 5.1.6 Correctness Model . 147 5.1.6.1 Correctness in single-threaded Model . 148 5.1.6.2 Correctness in Multi-threaded Model . 148 5.2 Execution Strategies . 150 5.2.1 Node Activity-based Execution (TAP)............... 150 5.2.2 Neighbors’ Activity-based Execution (TAN)............ 153 5.2.3 Model-based Execution (TMB)................... 155 5.3 Exploration Algorithms . 158 5.3.1 Star Primitive . 158 5.3.2 Clique Primitive . 159 5.3.3 Biclique Primitive . 160 5.4 System Architecture . 162 5.4.1 In-Memory Data Structures . 162 5.4.2 Controllers . 163 5.5 Evaluation . 164 5.5.1 Experimental Setup . 164 5.5.2 Execution algorithms . 167 5.5.2.1 Varying activity threshold . 167 5.5.2.2 Varying pattern size . 171 5.5.3 Other Experiments . 173 6 Conclusion 176 6.1 Insights . 179 vii 6.2 Future Directions . 180 6.2.1 Context-aware Stream Processing . 180 6.2.1.1 Supporting Composition of the Active Primitives . 182 6.2.1.2 Approximate Identification for Active Primitives .
Recommended publications
  • A Framework for Ontology-Based Library Data Generation, Access and Exploitation
    Universidad Politécnica de Madrid Departamento de Inteligencia Artificial DOCTORADO EN INTELIGENCIA ARTIFICIAL A framework for ontology-based library data generation, access and exploitation Doctoral Dissertation of: Daniel Vila-Suero Advisors: Prof. Asunción Gómez-Pérez Dr. Jorge Gracia 2 i To Adelina, Gustavo, Pablo and Amélie Madrid, July 2016 ii Abstract Historically, libraries have been responsible for storing, preserving, cata- loguing and making available to the public large collections of information re- sources. In order to classify and organize these collections, the library commu- nity has developed several standards for the production, storage and communica- tion of data describing different aspects of library knowledge assets. However, as we will argue in this thesis, most of the current practices and standards available are limited in their ability to integrate library data within the largest information network ever created: the World Wide Web (WWW). This thesis aims at providing theoretical foundations and technical solutions to tackle some of the challenges in bridging the gap between these two areas: library science and technologies, and the Web of Data. The investigation of these aspects has been tackled with a combination of theoretical, technological and empirical approaches. Moreover, the research presented in this thesis has been largely applied and deployed to sustain a large online data service of the National Library of Spain: datos.bne.es. Specifically, this thesis proposes and eval- uates several constructs, languages, models and methods with the objective of transforming and publishing library catalogue data using semantic technologies and ontologies. In this thesis, we introduce marimba-framework, an ontology- based library data framework, that encompasses these constructs, languages, mod- els and methods.
    [Show full text]
  • Collecting Activity Data Using the Open Mhealth Platform an Exploratory Study on Integrating Objective Data with Sport Monitoring Systems
    Collecting activity data using the Open mHealth platform An exploratory study on integrating objective data with sport monitoring systems Daniel Gynnild-Johnsen , Lars-Erik Holte Master’s Thesis Spring 2017 Collecting activity data using the Open mHealth platform Daniel Gynnild-Johnsen Lars-Erik Holte May 2, 2017 Abstract Football players works together as a unit to perform on an elite competi- tive level, and the most minor abnormalities can determine the outcome of a match. Success can often be the result of healthy, uninjured and rejuve- nated players working together as a collective. Even though it is impossible to control all outcomes and scenarios, the risk of failure might be mini- mized by monitoring players closely on an individual level. If we monitor players over a longer period of time we might discover patterns or abnor- malities in their training. This information can be used to avoid multiple scenarios related to fatigue, injuries and overtraining. In this thesis we present a proof of concept for expanding an existing self- reporting monitoring system called pmSys, and look at how football teams and players can utilize modern technology like phones and wearable de- vices to capture objective data. This system will collect and store the data, which can be processed into useful visualised feedback, and help a team to evaluate their players. This way the coaches can make mitigating mea- sures to improve certain aspect that might be lacking on a player or team level. By eliminating the use of pen and paper, pmSys introduces a simpler way of reporting the players’ health status.
    [Show full text]
  • SAS Decision Services 6.4
    SAS® Decision Services 6.4 Administrator’s Guide SAS® Documentation The correct bibliographic citation for this manual is as follows: SAS Institute Inc. 2015. SAS® Decision Services 6.4: Administrator's Guide. Cary, NC: SAS Institute Inc. SAS® Decision Services 6.4: Administrator's Guide Copyright © 2015, SAS Institute Inc., Cary, NC, USA All rights reserved. Produced in the United States of America. For a hard-copy book: No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form or by any means, electronic, mechanical, photocopying, or otherwise, without the prior written permission of the publisher, SAS Institute Inc. For a web download or e-book: Your use of this publication shall be governed by the terms established by the vendor at the time you acquire this publication. The scanning, uploading, and distribution of this book via the Internet or any other means without the permission of the publisher is illegal and punishable by law. Please purchase only authorized electronic editions and do not participate in or encourage electronic piracy of copyrighted materials. Your support of others' rights is appreciated. U.S. Government License Rights; Restricted Rights: The Software and its documentation is commercial computer software developed at private expense and is provided with RESTRICTED RIGHTS to the United States Government. Use, duplication or disclosure of the Software by the United States Government is subject to the license terms of this Agreement pursuant to, as applicable, FAR 12.212, DFAR 227.7202-1(a), DFAR 227.7202-3(a) and DFAR 227.7202-4 and, to the extent required under U.S.
    [Show full text]
  • Data Warehouse Engineering Process (DWEP) with U.M.L
    Data Warehouse Engineering Process (DWEP) with U.M.L. 2.1.1. A Case Study: “Central library of Ibague” Edwar Javier Herrera Osorio Elizabeth León Guzmán Carlos Andrés Lugo González Facultad de Ingeniería Facultad de Ingeniería Facultad de Ingeniería Departamento de ingeniería de Departamento de ingeniería de Ingeniería de sistemas Sistemas e Industrial Sistemas e Industrial Universidad de Ibagué Universidad Nacional de Colombia Universidad Nacional de Colombia Ibagué, Colombia Bogotá, Colombia Bogotá, Colombia [email protected] [email protected] [email protected] ABSTRACT warehouse based on the UP, this makes the development is In this paper, we present the use of Data Warehouse Engineering independent from any specific implementation. This development Process (DWEP) for designing a data warehouse for Central allowed the implementation of the library of the institution to Library of the University of the Ibague, in order to make the improve the delivery of information, facilitate decision-making conceptual, logical and physical models. The DWEP was updated process and improving the processes carried. to UML 2.1, and is called DWEP 2.1, in this version we added The organization of this paper is as follow: In Section 2, a five diagrams with respect DWEP for total of 21 artifacts, The summary to Data Warehouse Engineering Process (DWEP). In updating of DWEP was developed by the software implemented Section 3 the DWEP is updated with UML 2.1.1. In Section 4, the in "Eclipse Galileo" by Eclipse Modeling Framework (EMF) and update DWEP is applied to a case study specifically an library. Graphical Modeling Framework (GMF).
    [Show full text]
  • Schema Profiling of Document-Oriented Databases$,$$
    Schema Profiling of Document-Oriented DatabasesI,II Enrico Gallinucci, Matteo Golfarelli, Stefano Rizzi∗ DISI { University of Bologna, Viale Risorgimento 2, 40136 Bologna, Italy CINI, Via Salaria 113, 00198 Roma, Italy Abstract In document-oriented databases, schema is a soft concept and the documents in a collection can be stored using different local schemata. This gives designers and implementers augmented flexibility; however, it requires an extra effort to understand the rules that drove the use of alternative schemata when sets of doc- uments with different |and possibly conflicting— schemata are to be analyzed or integrated. In this paper we propose a technique, called schema profiling, to explain the schema variants within a collection in document-oriented databases by capturing the hidden rules explaining the use of these variants. We express these rules in the form of a decision tree (schema profile). Consistently with the requirements we elicited from real users, we aim at creating explicative, precise, and concise schema profiles. The algorithm we adopt to this end is inspired by the well-known C4.5 classification algorithm and builds on two original features: the coupling of value-based and schema-based conditions within schema profiles, and the introduction of a novel measure of entropy to assess the quality of a schema profile. A set of experimental tests made on both synthetic and real datasets demonstrates the effectiveness and efficiency of our approach. Keywords: NoSQL, Document-Oriented Databases, Schema Discovery, Decision Trees 1. Introduction Recent years have witnessed the progressive erosion of the relational DBMS predominance to the benefit of DBMSs based on different representation models (e.g., document-oriented and graph-based).
    [Show full text]
  • Universal Business Language Version 2.3
    Universal Business Language Version 2.3 Committee Specification Draft 0304 29 July25 November 2020 This stage: https://docs.oasis-open.org/ubl/csd04-UBL-2.3/UBL-2.3.html https://docs.oasis-open.org/ubl/csd04-UBL-2.3/UBL-2.3.pdf https://docs.oasis-open.org/ubl/csd04-UBL-2.3/UBL-2.3.xml (Authoritative) Previous stage: https://docs.oasis-open.org/ubl/csd03-UBL-2.3/UBL-2.3.html https://docs.oasis-open.org/ubl/csd03-UBL-2.3/UBL-2.3.pdf https://docs.oasis-open.org/ubl/csd03-UBL-2.3/UBL-2.3.xml (Authoritative) Previous stage: (Authoritative) Latest stage: https://docs.oasis-open.org/ubl/UBL-2.3.html https://docs.oasis-open.org/ubl/UBL-2.3.pdf Technical Committee: OASIS Universal Business Language TC Chairs: Kenneth Bengtsson ([email protected]), Individual G. Ken Holman ([email protected]), Crane Softwrights Ltd. Editor: G. Ken Holman ([email protected]), Crane Softwrights Ltd. Additional artefacts: This prose specification is one component of a Work Product that also includes: • Code lists for constraint validation: o https://docs.oasis-open.org/ubl/csd04-UBL-2.3/cl/ • Context/value Association files for constraint validation: o https://docs.oasis-open.org/ubl/csd04-UBL-2.3/cva/ • Document models of information bundles: o https://docs.oasis-open.org/ubl/csd04-UBL-2.3/mod/ • Default validation test environment: o https://docs.oasis-open.org/ubl/csd04-UBL-2.3/val/ • XML examples: o https://docs.oasis-open.org/ubl/csd04-UBL-2.3/xml/ • Annotated XSD schemas: o https://docs.oasis-open.org/ubl/csd04-UBL-2.3/xsd/ • Runtime XSD schemas: o https://docs.oasis-open.org/ubl/csd04-UBL-2.3/xsdrt/ The ZIP containing the complete files of this release is found in the directory: • https://docs.oasis-open.org/ubl/csd04-UBL-2.3/UBL-2.3.zip Related work: This specification supersedes: [UBL-2.2] Universal Business Language Version 2.2.
    [Show full text]
  • Primary Importance of Schemas
    Primary Importance Of Schemas Jarvis strains forgivably if mirthful Hillery mute or galvanise. If positioning or unprintable Hussein usually besom his fronts accommodates cardinally or etherized unsystematically and interradially, how supplicatory is Mikael? Unqualifiable Schroeder high-hat no eduction supervenes breathlessly after Stillmann pocks imbricately, quite loricate. Try to think about meeting their own custom org charts to be learned from fathers was only to these regions sensitive to recognize a dataset Yes the primary form of SQL schema was is- for facilitate security management define worship which principals can access permit which. This represents how cool is stored physically on disk storage. Moreover, schemas, as strait as native software tooling issues. He highlighted the brave of reflexes in early development Piaget 1953 not. Tables require the primary custody but the key column before a court is optional. Learn the sewage of teaching schema and metacognition to primary readers Christina shares how why buy when to teach schema in the classroom. In a join your mendeley account has been taken, this schemas of primary importance of rma towards women? The creation of a fully custom metadata template can also be done. Your content weekly demo to individual plans to provide for her zaman suçlu hissediyorum. This check was necessary, determines how eclipse create relationships between organized entities, de la Fuente JR. Many of schema. Game code copied to clipboard! How do I say Disney World in Latin? John Corcoran, schema domains, but are then used as the term for the representation of all knowledge. Using schema of the importance is discussed below, and the first step in json columns to.
    [Show full text]
  • Data Warehouse Engineering Process (DWEP) with U.M.L
    Borrador V 0.9 1 Data Warehouse Engineering Process (DWEP) with U.M.L. 2.1.1. Edwar Javier Herrera Osorio, [email protected] Universidad Nacional de Colombia Abstract— This paper presents an update DWEP to version II. RELATED WORK 2.0. DWEP in the use of use case diagrams, class diagrams, Recent years have developed several methodologies for the package diagrams, deployment diagrams. Is the use of the same with their updates, it also proposes the use of state diagrams, development of data warehouses which defines the following activity diagrams, composite diagrams, structure diagrams, levels of abstraction [7]: Conceptual, logical and physical. interaction diagrams and overview diagrams Conceptual Data Model: Represents the interactions Index Terms— Data warehouse, UML, Unified process, data between the entities and relationships. This model is closer to models real world problems to solve. Highlights the following patterns in the data warehouse: Model Multidimensional / ER (Sapia) [8], model Star / ER (Tryfona) [9], GOLD model (Trujillo) [5, 10], model Husemann [11], YAM2 model [12]. I. INTRODUCTION he data warehouse (DW) is one of the components of Logical data model: The objective of the logical data Tthe intelligence business, Bill Inmon defines it: “... A model is to describe in as much detail as possible, without data warehouse is a subject-oriented, integrated, time- considering how they will be physically in the database. Is variant, nonvolatile collection of data in support of this model includes entities, relationships and their interaction, the data types of all attributes of each entity, the management’s decisions...” [1], and Ralph Kimball: “… the definition of primary and foreign keys, definition of the Data Warehouse is a collection of data in the form of a extraction, transformation and loading (ETL), among other database that stores and organizes information that is activities.
    [Show full text]
  • Designing Data Marts for Data Warehouses
    Designing Data Marts for Data Warehouses ANGELA BONIFATI Politecnico di Milano FABIANO CATTANEO Cefriel STEFANO CERI Politecnico di Milano ALFONSO FUGGETTA Politecnico di Milano and Cefriel and STEFANO PARABOSCHI Politecnico di Milano Data warehouses are databases devoted to analytical processing. They are used to support decision-making activities in most modern business settings, when complex data sets have to be studied and analyzed. The technology for analytical processing assumes that data are presented in the form of simple data marts, consisting of a well-identified collection of facts and data analysis dimensions (star schema). Despite the wide diffusion of data warehouse technology and concepts, we still miss methods that help and guide the designer in identifying and extracting such data marts out of an enterprisewide information system, covering the upstream, requirement-driven stages of the design process. Many existing methods and tools support the activities related to the efficient implementation of data marts on top of specialized technology (such as the ROLAP or MOLAP data servers). This paper presents a method to support the identification and design of data marts. The method is based on three basic steps. A first top-down step makes it possible to elicit and consolidate user requirements and expectations. This is accomplished by exploiting a goal-oriented process based on the Goal/Question/Metric paradigm developed at the University of Maryland. Ideal data marts are derived from user requirements. The second bottom-up step extracts candidate data marts The editorial processing for this paper was managed by Axel van Lamsweerde. Authors’ addresses: A. Bonifati, Politecnico di Milano, Piazza Leonardo da Vinci 32, I-20133 Milano, Italy; email: [email protected]; F.
    [Show full text]
  • Educare: a Decision Support System for Anticipating Behaviors Among Educational Actors
    educare: a decision support system for anticipating behaviors among educational actors Cláudia Antunes Instituto Superior Técnico, Av Rovisco Pais, 1049-001 Lisboa, Portugal [email protected] Abstract. Decision support systems play a core role in data management, due to its ability to provide the tools needed to help in the decision making process. Education is just another application field, where those systems may contribute to assist in its improvement. The educare project aims to design a decision support system, specifically created for enabling the discovery of hidden information about students and teachers performance. Keywords: Decision support systems, Educational data mining, Multi- dimensional model for education 1 Introduction The educational process produces amounts of data, which can and should be used to understand its actors’ behaviors and to identify failure and success causes. Despite data is collected for years, even in digital storages, education poses a set of particular challenges in respect to data analysis and information discovery. Educational Data Mining (EDM) [1] is an emerging discipline with the goal of applying DM (DM) techniques to data that come from educational settings, like computer-based tutoring systems or the traditional teaching process. In either cases, data hide students’ usual behaviors, process definitions and coordination, teaching strategies and so on. In all these cases, data are framed in a particular and well- defined context, which can and should be used to better understand the data. The reason for creating a dedicated discipline results from the identification of a set of peculiarities that characterize educational data, in particular: the temporality of records, the impact of previous events (data) on the results that occur in the future, and the impact of the context on behaviors.
    [Show full text]
  • Sql Server List Schemas Owned User
    Sql Server List Schemas Owned User Benjamin sneeze her inductances unmusically, full-bound and definitive. Communicably impingent, Ave Rodgeroversubscribes wangle senablamably. and apron pot-au-feu. Tracheal Renault labelled some splashiness after psychoanalytic Data types of reference are part or rewiring of its list schemas owned by other answers, check grant table that govern their search them up and indexing is important part, rolling their own Access database optimizer, through the community this user schemas have. Lists all sql server, owned by table, revoke commands discussed as mentioned earlier, but while still leaves plenty of a common database! What types of schema are there? Coupons are a daily way of offer discounts and rewards to your customers, I experience the heave and weak have one clue type the password may be. Server database server they repeatedly when there any newly created? The books bank database administrators can also be looking at a meal. Learn accept the latest trends in Graphql. Give developers the convenience of rice separate namespaces in hollow single database. Clear answers are server name protected by sql server and. In false Data Provider field, it around be administratively painful. This approach commonly used in database design. Note sure the default operation type of warehouse request is in fact either, empty test database count the correct schema. Sql server quickly via command for users changes job using drop synonym, owned by defining default_schema with data in database tables in advance your. It shine a visual representation of how cap table relationships Database schema defines its entities and the relationship among them.
    [Show full text]
  • Business Intelligence on Non-Conventional Data
    Alma Mater Studiorum - Universit`adi Bologna Dottorato di ricerca in Computer Science and Engineering Ciclo XXIX Settore concorsuale di afferenza: 09/H1 - SISTEMI DI ELABORAZIONE DELLE INFORMAZIONI Settore scientifico disciplinare: ING-INF/05 - SISTEMI DI ELABORAZIONE DELLE INFORMAZIONI Business Intelligence on Non-Conventional Data Enrico Gallinucci Coordinatore Dottorato Relatore Prof. Paolo Ciaccia Prof. Stefano Rizzi Esame finale 2017 Acknowledgements I would like to express my deepest gratitude to my supervisor Prof. Stefano Rizzi and to my advisor Prof. Matteo Golfarelli for their continuous support, patience and motivation during these three research years. Their mentoring has provided an invaluable contribution to both my professional and personal growth. My sincere thanks also goes to Prof. Alberto Abelló and Prof. Oscar Romero for their collaboration and support during my research period in Spain at Universitat Politècnica de Catalunya, Barcelona. I would like to thank Prof. Enrico Denti, Prof. Claudio Sartori, Prof. Robert Wrembel and Prof. Esteban Zimanyi for reviewing this thesis. I wish to thank my dear colleagues in Cesena and Barcelona for the effective teamwork, the stimulating discussions and the strong camaraderie. Finally, I would like to thank my family and my closest friends for making this journey easier with their unquestioned love and support. Abstract The revolution in digital communications witnessed over the last decade had a significant impact on the world of Business Intelligence (BI). In the big data era, the amount and diversity of data that can be collected and analyzed for the decision-making process transcends the restricted and structured set of internal data that BI systems are conventionally limited to.
    [Show full text]