New Directions for Knowledge Representation on the Semantic Web
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
A Comparative Evaluation of Geospatial Semantic Web Frameworks for Cultural Heritage
heritage Article A Comparative Evaluation of Geospatial Semantic Web Frameworks for Cultural Heritage Ikrom Nishanbaev 1,* , Erik Champion 1,2,3 and David A. McMeekin 4,5 1 School of Media, Creative Arts, and Social Inquiry, Curtin University, Perth, WA 6845, Australia; [email protected] 2 Honorary Research Professor, CDHR, Sir Roland Wilson Building, 120 McCoy Circuit, Acton 2601, Australia 3 Honorary Research Fellow, School of Social Sciences, FABLE, University of Western Australia, 35 Stirling Highway, Perth, WA 6907, Australia 4 School of Earth and Planetary Sciences, Curtin University, Perth, WA 6845, Australia; [email protected] 5 School of Electrical Engineering, Computing and Mathematical Sciences, Curtin University, Perth, WA 6845, Australia * Correspondence: [email protected] Received: 14 July 2020; Accepted: 4 August 2020; Published: 12 August 2020 Abstract: Recently, many Resource Description Framework (RDF) data generation tools have been developed to convert geospatial and non-geospatial data into RDF data. Furthermore, there are several interlinking frameworks that find semantically equivalent geospatial resources in related RDF data sources. However, many existing Linked Open Data sources are currently sparsely interlinked. Also, many RDF generation and interlinking frameworks require a solid knowledge of Semantic Web and Geospatial Semantic Web concepts to successfully deploy them. This article comparatively evaluates features and functionality of the current state-of-the-art geospatial RDF generation tools and interlinking frameworks. This evaluation is specifically performed for cultural heritage researchers and professionals who have limited expertise in computer programming. Hence, a set of criteria has been defined to facilitate the selection of tools and frameworks. -
One Knowledge Graph to Rule Them All? Analyzing the Differences Between Dbpedia, YAGO, Wikidata & Co
One Knowledge Graph to Rule them All? Analyzing the Differences between DBpedia, YAGO, Wikidata & co. Daniel Ringler and Heiko Paulheim University of Mannheim, Data and Web Science Group Abstract. Public Knowledge Graphs (KGs) on the Web are consid- ered a valuable asset for developing intelligent applications. They contain general knowledge which can be used, e.g., for improving data analyt- ics tools, text processing pipelines, or recommender systems. While the large players, e.g., DBpedia, YAGO, or Wikidata, are often considered similar in nature and coverage, there are, in fact, quite a few differences. In this paper, we quantify those differences, and identify the overlapping and the complementary parts of public KGs. From those considerations, we can conclude that the KGs are hardly interchangeable, and that each of them has its strenghts and weaknesses when it comes to applications in different domains. 1 Knowledge Graphs on the Web The term \Knowledge Graph" was coined by Google when they introduced their knowledge graph as a backbone of a new Web search strategy in 2012, i.e., moving from pure text processing to a more symbolic representation of knowledge, using the slogan \things, not strings"1. Various public knowledge graphs are available on the Web, including DB- pedia [3] and YAGO [9], both of which are created by extracting information from Wikipedia (the latter exploiting WordNet on top), the community edited Wikidata [10], which imports other datasets, e.g., from national libraries2, as well as from the discontinued Freebase [7], the expert curated OpenCyc [4], and NELL [1], which exploits pattern-based knowledge extraction from a large Web corpus. -
Wikipedia Knowledge Graph with Deepdive
The Workshops of the Tenth International AAAI Conference on Web and Social Media Wiki: Technical Report WS-16-17 Wikipedia Knowledge Graph with DeepDive Thomas Palomares Youssef Ahres [email protected] [email protected] Juhana Kangaspunta Christopher Re´ [email protected] [email protected] Abstract This paper is organized as follows: first, we review the related work and give a general overview of DeepDive. Sec- Despite the tremendous amount of information on Wikipedia, ond, starting from the data preprocessing, we detail the gen- only a very small amount is structured. Most of the informa- eral methodology used. Then, we detail two applications tion is embedded in unstructured text and extracting it is a non trivial challenge. In this paper, we propose a full pipeline that follow this pipeline along with their specific challenges built on top of DeepDive to successfully extract meaningful and solutions. Finally, we report the results of these applica- relations from the Wikipedia text corpus. We evaluated the tions and discuss the next steps to continue populating Wiki- system by extracting company-founders and family relations data and improve the current system to extract more relations from the text. As a result, we extracted more than 140,000 with a high precision. distinct relations with an average precision above 90%. Background & Related Work Introduction Until recently, populating the large knowledge bases relied on direct contributions from human volunteers as well With the perpetual growth of web usage, the amount as integration of existing repositories such as Wikipedia of unstructured data grows exponentially. Extract- info boxes. These methods are limited by the available ing facts and assertions to store them in a struc- structured data and by human power. -
Knowledge Graphs on the Web – an Overview Arxiv:2003.00719V3 [Cs
January 2020 Knowledge Graphs on the Web – an Overview Nicolas HEIST, Sven HERTLING, Daniel RINGLER, and Heiko PAULHEIM Data and Web Science Group, University of Mannheim, Germany Abstract. Knowledge Graphs are an emerging form of knowledge representation. While Google coined the term Knowledge Graph first and promoted it as a means to improve their search results, they are used in many applications today. In a knowl- edge graph, entities in the real world and/or a business domain (e.g., people, places, or events) are represented as nodes, which are connected by edges representing the relations between those entities. While companies such as Google, Microsoft, and Facebook have their own, non-public knowledge graphs, there is also a larger body of publicly available knowledge graphs, such as DBpedia or Wikidata. In this chap- ter, we provide an overview and comparison of those publicly available knowledge graphs, and give insights into their contents, size, coverage, and overlap. Keywords. Knowledge Graph, Linked Data, Semantic Web, Profiling 1. Introduction Knowledge Graphs are increasingly used as means to represent knowledge. Due to their versatile means of representation, they can be used to integrate different heterogeneous data sources, both within as well as across organizations. [8,9] Besides such domain-specific knowledge graphs which are typically developed for specific domains and/or use cases, there are also public, cross-domain knowledge graphs encoding common knowledge, such as DBpedia, Wikidata, or YAGO. [33] Such knowl- edge graphs may be used, e.g., for automatically enriching data with background knowl- arXiv:2003.00719v3 [cs.AI] 12 Mar 2020 edge to be used in knowledge-intensive downstream applications. -
Tractatus Logico-Philosophicus</Em>
University of South Florida Scholar Commons Graduate Theses and Dissertations Graduate School 8-6-2008 Three Wittgensteins: Interpreting the Tractatus Logico-Philosophicus Thomas J. Brommage Jr. University of South Florida Follow this and additional works at: https://scholarcommons.usf.edu/etd Part of the American Studies Commons Scholar Commons Citation Brommage, Thomas J. Jr., "Three Wittgensteins: Interpreting the Tractatus Logico-Philosophicus" (2008). Graduate Theses and Dissertations. https://scholarcommons.usf.edu/etd/149 This Dissertation is brought to you for free and open access by the Graduate School at Scholar Commons. It has been accepted for inclusion in Graduate Theses and Dissertations by an authorized administrator of Scholar Commons. For more information, please contact [email protected]. Three Wittgensteins: Interpreting the Tractatus Logico-Philosophicus by Thomas J. Brommage, Jr. A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy Department of Philosophy College of Arts and Sciences University of South Florida Co-Major Professor: Kwasi Wiredu, B.Phil. Co-Major Professor: Stephen P. Turner, Ph.D. Charles B. Guignon, Ph.D. Richard N. Manning, J. D., Ph.D. Joanne B. Waugh, Ph.D. Date of Approval: August 6, 2008 Keywords: Wittgenstein, Tractatus Logico-Philosophicus, logical empiricism, resolute reading, metaphysics © Copyright 2008 , Thomas J. Brommage, Jr. Acknowledgments There are many people whom have helped me along the way. My most prominent debts include Ray Langely, Billy Joe Lucas, and Mary T. Clark, who trained me in philosophy at Manhattanville College; and also to Joanne Waugh, Stephen Turner, Kwasi Wiredu and Cahrles Guignon, all of whom have nurtured my love for the philosophy of language. -
First Order Logic
Artificial Intelligence CS 165A Mar 10, 2020 Instructor: Prof. Yu-Xiang Wang ® First Order Logic 1 Recap: KB Agents True sentences TELL Knowledge Base Domain specific content; facts ASK Inference engine Domain independent algorithms; can deduce new facts from the KB • Need a formal logic system to work • Need a data structure to represent known facts • Need an algorithm to answer ASK questions 2 Recap: syntax and semantics • Two components of a logic system • Syntax --- How to construct sentences – The symbols – The operators that connect symbols together – A precedence ordering • Semantics --- Rules the assignment of sentences to truth – For every possible worlds (or “models” in logic jargon) – The truth table is a semantics 3 Recap: Entailment Sentences Sentence ENTAILS Representation Semantics Semantics World Facts Fact FOLLOWS A is entailed by B, if A is true in all possiBle worlds consistent with B under the semantics. 4 Recap: Inference procedure • Inference procedure – Rules (algorithms) that we apply (often recursively) to derive truth from other truth. – Could be specific to a particular set of semantics, a particular realization of the world. • Soundness and completeness of an inference procedure – Soundness: All truth discovered are valid. – Completeness: All truth that are entailed can be discovered. 5 Recap: Propositional Logic • Syntax:Syntax – True, false, propositional symbols – ( ) , ¬ (not), Ù (and), Ú (or), Þ (implies), Û (equivalent) • Semantics: – Assigning values to the variables. Each row is a “model”. • Inference rules: – Modus Pronens etc. Most important: Resolution 6 Recap: Propositional logic agent • Representation: Conjunctive Normal Forms – Represent them in a data structure: a list, a heap, a tree? – Efficient TELL operation • Inference: Solve ASK question – Use “Resolution” only on CNFs is Sound and Complete. -
Knowledge Graph Identification
Knowledge Graph Identification Jay Pujara1, Hui Miao1, Lise Getoor1, and William Cohen2 1 Dept of Computer Science, University of Maryland, College Park, MD 20742 fjay,hui,[email protected] 2 Machine Learning Dept, Carnegie Mellon University, Pittsburgh, PA 15213 [email protected] Abstract. Large-scale information processing systems are able to ex- tract massive collections of interrelated facts, but unfortunately trans- forming these candidate facts into useful knowledge is a formidable chal- lenge. In this paper, we show how uncertain extractions about entities and their relations can be transformed into a knowledge graph. The ex- tractions form an extraction graph and we refer to the task of removing noise, inferring missing information, and determining which candidate facts should be included into a knowledge graph as knowledge graph identification. In order to perform this task, we must reason jointly about candidate facts and their associated extraction confidences, identify co- referent entities, and incorporate ontological constraints. Our proposed approach uses probabilistic soft logic (PSL), a recently introduced prob- abilistic modeling framework which easily scales to millions of facts. We demonstrate the power of our method on a synthetic Linked Data corpus derived from the MusicBrainz music community and a real-world set of extractions from the NELL project containing over 1M extractions and 70K ontological relations. We show that compared to existing methods, our approach is able to achieve improved AUC and F1 with significantly lower running time. 1 Introduction The web is a vast repository of knowledge, but automatically extracting that knowledge at scale has proven to be a formidable challenge. -
Arxiv:1910.13561V1 [Cs.LG] 29 Oct 2019 E-Mail: [email protected] M
Noname manuscript No. (will be inserted by the editor) A Heuristically Modified FP-Tree for Ontology Learning with Applications in Education Safwan Shatnawi · Mohamed Medhat Gaber ∗ · Mihaela Cocea Received: date / Accepted: date Abstract We propose a heuristically modified FP-Tree for ontology learning from text. Unlike previous research, for concept extraction, we use a regular expression parser approach widely adopted in compiler construction, i.e., deterministic finite automata (DFA). Thus, the concepts are extracted from unstructured documents. For ontology learning, we use a frequent pattern mining approach and employ a rule mining heuristic function to enhance its quality. This process does not rely on predefined lexico-syntactic patterns, thus, it is applicable for different subjects. We employ the ontology in a question-answering system for students' content-related questions. For validation, we used textbook questions/answers and questions from online course forums. Subject experts rated the quality of the system's answers on a subset of questions and their ratings were used to identify the most appropriate automatic semantic text similarity metric to use as a validation metric for all answers. The Latent Semantic Analysis was identified as the closest to the experts' ratings. We compared the use of our ontology with the use of Text2Onto for the question-answering system and found that with our ontology 80% of the questions were answered, while with Text2Onto only 28.4% were answered, thanks to the finer grained hierarchy our approach is able to produce. Keywords Ontologies · Frequent pattern mining · Ontology learning · Question answering · MOOCs S. Shatnawi College of Applied Studies, University of Bahrain, Sakhair Campus, Zallaq, Bahrain E-mail: [email protected] M. -
Building Dialogue Structure from Discourse Tree of a Question
The Workshops of the Thirty-Second AAAI Conference on Artificial Intelligence Building Dialogue Structure from Discourse Tree of a Question Boris Galitsky Oracle Corp. Redwood Shores CA USA [email protected] Abstract ed, chat bot’s capability to maintain the cohesive flow, We propose a reasoning-based approach to a dialogue man- style and merits of conversation is an underexplored area. agement for a customer support chat bot. To build a dia- When a question is detailed and includes multiple sen- logue scenario, we analyze the discourse tree (DT) of an ini- tences, there are certain expectations concerning the style tial query of a customer support dialogue that is frequently of an answer. Although a topical agreement between ques- complex and multi-sentence. We then enforce what we call tions and answers have been extensively addressed, a cor- complementarity relation between DT of the initial query respondence in style and suitability for the given step of a and that of the answers, requests and responses. The chat bot finds answers, which are not only relevant by topic but dialogue between questions and answers has not been thor- also suitable for a given step of a conversation and match oughly explored. In this study we focus on assessment of the question by style, argumentation patterns, communica- cohesiveness of question/answer (Q/A) flow, which is im- tion means, experience level and other domain-independent portant for a chat bots supporting longer conversation. attributes. We evaluate a performance of proposed algo- When an answer is in a style disagreement with a question, rithm in car repair domain and observe a 5 to 10% im- a user can find this answer inappropriate even when a topi- provement for single and three-step dialogues respectively, in comparison with baseline approaches to dialogue man- cal relevance is high. -
Frege and the Logic of Sense and Reference
FREGE AND THE LOGIC OF SENSE AND REFERENCE Kevin C. Klement Routledge New York & London Published in 2002 by Routledge 29 West 35th Street New York, NY 10001 Published in Great Britain by Routledge 11 New Fetter Lane London EC4P 4EE Routledge is an imprint of the Taylor & Francis Group Printed in the United States of America on acid-free paper. Copyright © 2002 by Kevin C. Klement All rights reserved. No part of this book may be reprinted or reproduced or utilized in any form or by any electronic, mechanical or other means, now known or hereafter invented, including photocopying and recording, or in any infomration storage or retrieval system, without permission in writing from the publisher. 10 9 8 7 6 5 4 3 2 1 Library of Congress Cataloging-in-Publication Data Klement, Kevin C., 1974– Frege and the logic of sense and reference / by Kevin Klement. p. cm — (Studies in philosophy) Includes bibliographical references and index ISBN 0-415-93790-6 1. Frege, Gottlob, 1848–1925. 2. Sense (Philosophy) 3. Reference (Philosophy) I. Title II. Studies in philosophy (New York, N. Y.) B3245.F24 K54 2001 12'.68'092—dc21 2001048169 Contents Page Preface ix Abbreviations xiii 1. The Need for a Logical Calculus for the Theory of Sinn and Bedeutung 3 Introduction 3 Frege’s Project: Logicism and the Notion of Begriffsschrift 4 The Theory of Sinn and Bedeutung 8 The Limitations of the Begriffsschrift 14 Filling the Gap 21 2. The Logic of the Grundgesetze 25 Logical Language and the Content of Logic 25 Functionality and Predication 28 Quantifiers and Gothic Letters 32 Roman Letters: An Alternative Notation for Generality 38 Value-Ranges and Extensions of Concepts 42 The Syntactic Rules of the Begriffsschrift 44 The Axiomatization of Frege’s System 49 Responses to the Paradox 56 v vi Contents 3. -
New Generation of Social Networks Based on Semantic Web Technologies: the Importance of Social Data Portability
Workshop on Adaptation and Personalization for Web 2.0, UMAP'09, June 22-26, 2009 New Generation of Social Networks Based on Semantic Web Technologies: the Importance of Social Data Portability Liana Razmerita1, Martynas Jusevičius2, Rokas Firantas2 Copenhagen Business School, Denmark1 IT University of Copenhagen, Copenhagen, Denmark2 [email protected], [email protected], [email protected] Abstract. This article investigates several well-known social network applications such as Last.fm, Flickr and identifies social data portability as one of the main technical issues that need to be addressed in the future. We argue that this issue can be addressed by building social networks as Semantic Web applications with FOAF, SIOC, and Linked Data technologies, and prove it by implementing a prototype application using Java and core Semantic Web standards. Furthermore, the developed prototype shows how features from semantic websites such as Freebase and DBpedia can be reused in social applications and lead to more relevant content and stronger social connections. 1 Introduction Social networking sites are developing at a very fast pace, attracting more and more users. Their application domain can be music sharing, photos sharing, videos sharing, bookmarks sharing, professional networks, and others. Despite their tremendous success social networking websites have a number of limitations that are identified and discussed within this article. Most of the social networking applications are “walled websites” and the “online communities are like islands in a sea”[1]. Lack of interoperability between data in different social networks applications limits access to relevant content available on different social networking sites, and limits the integration and reuse of available data and information. -
How Much Is a Triple? Estimating the Cost of Knowledge Graph Creation
How much is a Triple? Estimating the Cost of Knowledge Graph Creation Heiko Paulheim Data and Web Science Group, University of Mannheim, Germany [email protected] Abstract. Knowledge graphs are used in various applications and have been widely analyzed. A question that is not very well researched is: what is the price of their production? In this paper, we propose ways to estimate the cost of those knowledge graphs. We show that the cost of manually curating a triple is between $2 and $6, and that the cost for automatically created knowledge graphs is by a factor of 15 to 250 cheaper (i.e., 1g to 15g per statement). Furthermore, we advocate for taking cost into account as an evaluation metric, showing the correspon- dence between cost per triple and semantic validity as an example. Keywords: Knowledge Graphs, Cost Estimation, Automation 1 Estimating the Cost of Knowledge Graphs With the increasing attention larger scale knowledge graphs, such as DBpedia [5], YAGO [12] and the like, have drawn towards themselves, they have been inspected under various angles, such as as size, overlap [11], or quality [3]. How- ever, one dimension that is underrepresented in those analyses is cost, i.e., the prize of creating the knowledge graphs. 1.1 Manual Curation: Cyc and Freebase For manually created knowledge graphs, we have to estimate the effort of pro- viding the statements directly. Cyc [6] is one of the earliest general purpose knowledge graphs, and, at the same time, the one for which the development effort is known. At a 2017 con- ference, Douglas Lenat, the inventor of Cyc, denoted the cost of creation of Cyc at $120M.1 In the same presentation, Lenat states that Cyc consists of 21M assertions, which makes a cost of $5.71 per statement.