Gstore: a Graph-Based SPARQL Query Engine

Total Page:16

File Type:pdf, Size:1020Kb

Gstore: a Graph-Based SPARQL Query Engine Noname manuscript No. (will be inserted by the editor) gStore: A Graph-based SPARQL Query Engine Lei Zou · M. Tamer Ozsu¨ · Lei Chen · Xuchuan Shen · Ruizhe Huang · Dongyan Zhao the date of receipt and acceptance should be inserted later Abstract We address efficient processing of SPARQL 1 Introduction queries over RDF datasets. The proposed techniques, incorporated into the gStore system, handle, in a uni- The RDF (Resource Description Framework) data model form and scalable manner, SPARQL queries with wild- was proposed for modeling Web objects as part of devel- cards and aggregate operators over dynamic RDF datasets. oping the semantic web. Its use in various applications Our approach is graph-based. We store RDF data as a is increasing. large graph, and also represent a SPARQL query as a A RDF data set is a collection of (subject, property, query graph. Thus the query answering problem is con- object) triples denoted as hs; p; oi. A running example verted into a subgraph matching problem. To achieve is given in Figure 1a. In order to query RDF reposito- efficient and scalable query processing, we develop an ries, SPARQL query language [23] has been proposed index, together with effective pruning rules and efficient by W3C. An example query that retrieves the names search algorithms. We propose techniques that use this of individuals who were born on February 12, 1809 and infrastructure to answer aggregation queries. We also who died on April 15, 1865 can be specified by the fol- propose an effective maintenance algorithm to handle lowing SPARQL query (Q1): online updates over RDF repositories. Extensive exper- SELECT ?name WHERE iments confirm the efficiency and effectiveness of our f?m <hasName> ?name . solutions. ?m <bornOnDate> ` `1809 −02 −12 ' '. ?m <diedOnDate> ` `1865 −04 −15 ' '.g Although RDF data management has been studied Extended version of paper \gStore: Answering SPARQL over the past decade, most early solutions do not scale Queries via Subgraph Matching" that was presented at 2011 to large RDF repositories and cannot answer complex VLDB Conference. queries efficiently. For example, early systems such as Lei Zou, Xuchuan Shen, Ruizhe Huang, Dongyan Zhao Jena [31], Yars2 [14] and Sesame 2.0 [6], do not work Institute of Computer Science and Technology, well over large RDF datasets. More recent works (e.g., Peking University, Beijing, China Tel.: +86-10-82529643 [2,20,33]) as well as systems, such as RDF-3x [19], x- E-mail: fzoulei,shenxuchuan,huangruizhe,[email protected] RDF-3x [22], Hexastore [30] and SW-store [1], are de- M. Tamer Ozsu¨ signed to address scalability over large data sets. How- David R. Cheriton School of Computer Science ever, none of these address scalability along with the University of Waterloo, Waterloo, Canada following real requirements of RDF applications: Tel.: +1-519-888-4043 E-mail: [email protected] { SPARQL queries with wildcards. Similar to SQL and Lei Chen XPath counterparts, the wildcard SPARQL queries Department of Computer Science and Engineering, Hong Kong University of Science and Technology, enable users to specify more flexible query criteria in Hong Kong, China real-life applications where users may not have full Tel.: +852-23586980 knowledge about a query object. For example, we E-mail: [email protected] may know that a person was born in 1976 in a city 2 Subject Property Object y:Abraham Lincoln hasName \Abraham Lincoln" y:Abraham Lincoln bornOnDate \1809-02-12" y:Abraham Lincoln diedOnDate \1865-04-15" 030 031 y:Abraham Lincoln bornIn y:Hodgenville KY \1926-07-01" \Female" y:Abraham Lincoln diedIn y:Washington DC bornOnDate gender y:Abraham Lincoln title \President" 009 029 diedOnDate hasName 032 y:Abraham Lincoln gender \Male" \1962-08-05" y:Marilyn Monroe \Marilyn Monroe" y:Washington DC hasName \Washington D.C." y:Washington DC foundingYear \1790" 010 013 014 y:Hodgenville KY hasName \Hodgenville" \Abraham Lincoln" \President" \Male" y:United States hasName \United States" hasName title gender 001 003 y:United States hasCapital y:Washington DC 011 bornOnDate bornIn hasName 017 y:United States foundingYear \1776" \1809-02-12" y:Abraham Lincoln y:Hodgenville KY \Hodgenville" y:Reese Witherspoon bornOnDate \1976-03-22" diedOnDate 025 026 012 diedIn 019 \Franklin D. Roosevelt" \Male" y:Reese Witherspoon bornIn y:New Orleans LA \1865-04-15" 002 \1776" hasName gender y:Reese Witherspoon hasName \Reese Witherspoon" y:Washington D.C. 007 020 y:Franklin Roosevelt y:Reese Witherspoon gender \Female" foundYear foundingYear \1976-03-22" 015 hasName hasCapital y:Reese Witherspoon title \Actress" 016 004 title \1790" 027 bornOnDate 005 \Washington D.C." y:United States y:New Orleans LA foundingYear \1718" \President" y:New Orleans LA locatedIn y:United States y:Reese Witherspoon bornIn y:Franklin Roosevelt hasName \Franklin D. Roo- gender hasName 021 locatedIn locatedIn title hasName bornIn sevelt" \Female" 018 y:Franklin Roosevelt bornIn y:Hyde Park NY 022 023 006 \United States" 008 \Reese Witherspoon" y:Franklin Roosevelt title \President" \Actress" y:New Orleans LA y:Hyde Park NY y:Franklin Roosevelt gender \Male" foundingYear foundingYear y:Hyde Park NY foundingYear \1810" 024 028 y:Hyde Park NY locatedIn y:United States \1718" \1810" y:Marilyn Monroe gender \Female" y:Marilyn Monroe hasName \Marilyn Monroe" y:Marilyn Monroe bornOnDate \1926-07-01" y:Marilyn Monroe diedOnDate \1962-08-05" (a) RDF Triples (Prefix: y=http://en.wikipedia.org/wiki/) (b) RDF Graph G Fig. 1: RDF Graph that was founded in 1718, but we may not know the In this paper we describe gStore, which is a graph- exact birth date. In this case, we have to perform a based triple store that can answer the above discussed query with wildcards, as shown below (Q2): SPARQL queries over dynamic RDF data repositories. SELECT ?name WHERE In this context, answering a query is transformed into f?m <bornIn> ? c i t y . ?m <hasName> ?name . a subgraph matching problem. Specifically, we model ?m <bornOnDate> ?bd . an RDF dataset (a collection of triples) as a labeled, ? c i t y <foundingYear> ` ` 1 7 1 8 ' ' . directed multi-edge graph (RDF graph), where each FILTER(regex(str(?bd), ` `1976 ' ')) g vertex corresponds to a subject or an object. We also { Dynamic RDF repositories. RDF repositories are represent a given SPARQL query by a query graph, Q. not static, and are updated regularly. For example, Subgraph matching of the query graph Q over the RDF Yago and DBpedia datasets are continually expand- graph G provides the answer to the query. ing to include the newly extracted knowledge from For example, Figure 1b shows an RDF graph G cor- Wikipedia. The RDF data in social networks, such responding to RDF triples in Figure 1a. We formally as the FOAF project (foaf-project.org), are also fre- define an RDF graph in Definition 1. Note that, the quently updated to represent the individuals' chang- numbers above the boxes in Figure 1b are not vertex ing relationships. In order to support queries over labels, but vertex IDs that we introduce to simplify the such dynamic RDF datasets, query engines should description. The RDF graph does not have to be con- be able to handle frequent updates without much nected. A SPARQL query can also be represented as a maintenance overhead. directed labeled query graph Q (Definition 2). Figure 2 { Aggregate SPARQL queries. Few existing works and shows the query graph corresponding to the SPARQL SPARQL engines consider aggregate queries despite query Q2. Usually, query graph Q is a connected graph. their real-life importance. A typical aggregate SPARQL Otherwise, we can regard each connected component of query that groups all individuals by their titles, gen- Q as a separate query and perform them one by one. ders, and the founding year of their birth places, and reports the number of individuals in each group is We develop novel indexing and graph matching tech- niques rather than using existing ones. This is because shown below (Q3): the characteristics of an RDF graph are considerably SELECT ? t ?g ?y COUNT(?m) WHERE different from graphs typically considered in much of f?m <bornIn> ? c . ?m <t i t l e > ? t . ?m <gender> ?g . ? c <foundingYear> ?y . g the graph database research. First, the size of an RDF GROUP BY ? t ?g ?y graph (i.e., the number of vertices and edges) is larger 3 regex(str(?bd),\1976") discuss the maintenance of indexes (VS∗-tree and T- index) as RDF data get updated in Section 10. We ?name ?bd \1718" study our methods by experiments in Section 11. Sec- hasName bornOnDate foundingYear tion 12 concludes this paper. Some of the additional bornIn ?m ?city material supporting the main findings reported in the paper are included in an Online Supplement. Fig. 2: Query Graph of Q2 2 Related Work than what is considered in typical graph databases by orders of magnitude. Second, the cardinality of ver- Three approaches have been proposed to store and query tex and edge labels in an RDF graph is much larger RDF data: one giant triples table, clustered property than that in traditional graph databases. For example, tables, and vertically partitioned tables. a typical dataset (i.e., the AIDS dataset) used in the existing graph database work [25,32] has 10,000 data graphs, each with an average number of 20 vertices One giant triples table. The systems in this category and 25 edges. The total number of distinct vertex la- store RDF triples in a single three-column table where bels is 62. The total size of the dataset is about 5M columns correspond to subject, property, and object bytes. However, the Yago RDF graph has about 500M (as in Figure 1a) enabling them to manipulate all RDF vertices and the total size is about 3.1GB.
Recommended publications
  • Injection of Automatically Selected Dbpedia Subjects in Electronic
    Injection of Automatically Selected DBpedia Subjects in Electronic Medical Records to boost Hospitalization Prediction Raphaël Gazzotti, Catherine Faron Zucker, Fabien Gandon, Virginie Lacroix-Hugues, David Darmon To cite this version: Raphaël Gazzotti, Catherine Faron Zucker, Fabien Gandon, Virginie Lacroix-Hugues, David Darmon. Injection of Automatically Selected DBpedia Subjects in Electronic Medical Records to boost Hos- pitalization Prediction. SAC 2020 - 35th ACM/SIGAPP Symposium On Applied Computing, Mar 2020, Brno, Czech Republic. 10.1145/3341105.3373932. hal-02389918 HAL Id: hal-02389918 https://hal.archives-ouvertes.fr/hal-02389918 Submitted on 16 Dec 2019 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Injection of Automatically Selected DBpedia Subjects in Electronic Medical Records to boost Hospitalization Prediction Raphaël Gazzotti Catherine Faron-Zucker Fabien Gandon Université Côte d’Azur, Inria, CNRS, Université Côte d’Azur, Inria, CNRS, Inria, Université Côte d’Azur, CNRS, I3S, Sophia-Antipolis, France I3S, Sophia-Antipolis, France
    [Show full text]
  • Knowledge Graphs on the Web – an Overview Arxiv:2003.00719V3 [Cs
    January 2020 Knowledge Graphs on the Web – an Overview Nicolas HEIST, Sven HERTLING, Daniel RINGLER, and Heiko PAULHEIM Data and Web Science Group, University of Mannheim, Germany Abstract. Knowledge Graphs are an emerging form of knowledge representation. While Google coined the term Knowledge Graph first and promoted it as a means to improve their search results, they are used in many applications today. In a knowl- edge graph, entities in the real world and/or a business domain (e.g., people, places, or events) are represented as nodes, which are connected by edges representing the relations between those entities. While companies such as Google, Microsoft, and Facebook have their own, non-public knowledge graphs, there is also a larger body of publicly available knowledge graphs, such as DBpedia or Wikidata. In this chap- ter, we provide an overview and comparison of those publicly available knowledge graphs, and give insights into their contents, size, coverage, and overlap. Keywords. Knowledge Graph, Linked Data, Semantic Web, Profiling 1. Introduction Knowledge Graphs are increasingly used as means to represent knowledge. Due to their versatile means of representation, they can be used to integrate different heterogeneous data sources, both within as well as across organizations. [8,9] Besides such domain-specific knowledge graphs which are typically developed for specific domains and/or use cases, there are also public, cross-domain knowledge graphs encoding common knowledge, such as DBpedia, Wikidata, or YAGO. [33] Such knowl- edge graphs may be used, e.g., for automatically enriching data with background knowl- arXiv:2003.00719v3 [cs.AI] 12 Mar 2020 edge to be used in knowledge-intensive downstream applications.
    [Show full text]
  • A Question Answering System Using Graph-Pattern Association Rules (QAGPAR) on YAGO Knowledge Base
    A Question Answering System Using Graph-Pattern Association Rules (QAGPAR) On YAGO Knowledge Base Wahyudi 1,2 , Masayu Leylia Khodra 2, Ary Setijadi Prihatmanto 2, Carmadi Machbub 2 1 1) Faculty of Information Technology, University of Andalas Padang, Indonesia 2) School of Electrical Engineering and Informatics, Institut Teknologi Bandung Bandung, Indonesia [email protected], [email protected], [email protected] , [email protected] Abstract. A question answering system (QA System) was developed that uses graph-pattern association rules on the YAGO knowledge base. The answer as output of the system is provided based on a user question as input. If the answer is missing or unavailable in the database, then graph-pattern association rules are used to get the answer. The architecture of this question answering system is as follows: question classification, graph component generation, query generation, and query processing. The question answering system uses association graph patterns in a waterfall model. In this paper, the architecture of the system is described, specifically discussing its reasoning and performance capabilities. The results of this research is that rules with high confidence and correct logic produce correct answers, and vice versa. ACM CCS (2012) Classification: CCS → Computing methodologies → Artificial intelligence → Knowledge representation and reasoning → Reasoning about belief and knowledge Keywords: graph pattern, question answering system, system architecture 1. Introduction Knowledge bases provide information about a large variety of entities, such as people, countries, rivers, etc. Moreover, knowledge bases also contain facts related to these entities, e.g., who lives where, which river is located in which city [1].
    [Show full text]
  • YAGO - Yet Another Great Ontology YAGO: a Large Ontology from Wikipedia and Wordnet1
    YAGO - Yet Another Great Ontology YAGO: A Large Ontology from Wikipedia and WordNet1 Presentation by: Besnik Fetahu UdS February 22, 2012 1 Fabian M.Suchanek, Gjergji Kasneci, Gerhard Weikum Presentation by: Besnik Fetahu (UdS) YAGO - Yet Another Great Ontology February 22, 2012 1 / 30 1 Introduction & Background Ontology-Review Usage of Ontology Related Work 2 The YAGO Model Aims of YAGO Representation Models Semantics 3 Putting all at work Information Extraction Approaches Quality Control 4 Evaluation & Examples 5 Conclusions Presentation by: Besnik Fetahu (UdS) YAGO - Yet Another Great Ontology February 22, 2012 2 / 30 Introduction & Background 1 Introduction & Background Ontology-Review Usage of Ontology Related Work 2 The YAGO Model Aims of YAGO Representation Models Semantics 3 Putting all at work Information Extraction Approaches Quality Control 4 Evaluation & Examples 5 Conclusions Presentation by: Besnik Fetahu (UdS) YAGO - Yet Another Great Ontology February 22, 2012 3 / 30 Introduction & Background Ontology-Review Ontology - Review Knowledge represented as a set of concepts within a domain, and a relationship between those concepts.2 In general it has the following components: Entities Relations Domains Rules Axioms, etc. 2 http://en.wikipedia.org/wiki/Ontology (information science) Presentation by: Besnik Fetahu (UdS) YAGO - Yet Another Great Ontology February 22, 2012 4 / 30 Introduction & Background Ontology-Review Ontology - Review Knowledge represented as a set of concepts within a domain, and a relationship between those concepts.2 In general it has the following components: Entities Relations Domains Rules Axioms, etc. 2 http://en.wikipedia.org/wiki/Ontology (information science) Presentation by: Besnik Fetahu (UdS) YAGO - Yet Another Great Ontology February 22, 2012 4 / 30 Word Sense Disambiguation Wikipedia's categories, hyperlinks, and disambiguous articles, used to create a dataset of named entity (Bunescu R., and Pasca M., 2006).
    [Show full text]
  • A Note on General Statistics of Publicly Accessible Knowledge Bases
    A Note on General Statistics of Publicly Accessible Knowledge Bases Feixiang Wang, Yixiang Fang, Yan Song, Shuang Li, Xinyun Chen The Chinese University of Hong Kong, Shenzhen [email protected],{fangyixiang,songyan,lishuang,chenxinyun}@cuhk.edu.cn ABSTRACT bases, which include their numbers of objects, relations, object Knowledge bases are prevalent in various domains and have been types, relation types, etc. Table 1 lists the publicly accessible knowl- 2 widely used in a large number of real applications such as ap- edge bases that are reviewed in this paper . As shown in Table 1, plications in online encyclopedia, social media, biomedical fields, we group these knowledge bases into two categories, i.e., schema- bibliographical networks. Due to their great importance, knowl- rich and simple-schema knowledge bases. Herein, the schema of a edge bases have received much attention from both the academia knowledge base refers to a network describing all allowable rela- and industry community in recent years. In this paper, we provide tion types between entities types and also the properties of entities a summary of the general statistics of several open-source and [10]. For example, if an entity is “Taylor Swif”, then the schema publicly accessible knowledge bases, ranging from the number of defines the type of this entity (i.e., Person) and the properties ofthis objects, relations to the object types and the relation types. With entity in this type (i.e., her age, job, country, etc.). Therefore, the such statistics, this concise note can not only help researchers form schema-rich knowledge bases often refer to the knowledge bases a better and quick understanding of existing publicly accessible with a large number of object/relation types, while the simple- knowledge bases, but can also guide the general audience to use the schema knowledge bases only have a few object/relation types.
    [Show full text]
  • INFO/CS 4302 Web Informa>On Systems
    INFO/CS 4302 Web Informaon Systems FT 2012 Lecture 17: SKOS, SPARQL - Bernhard HasLhofer - PLan for today... • Context & Recap • SKOS • SPARQL • Quesons 2 CONTEXT & RECAP 3 4 SPARQL • A Language for describing knowLedge organizaon systems (taxonomies, thesauri, cLassificaon schemes) http://dbpedia.org/resource/ Category:1980s_horror_films dcterms:subject rdf:type skos:broader http://dbpedia.org/resource/ http://www.w3.org/2004/02/ The_Shining_(film) skos/core#Concept rdf:type http://dbpedia.org/resource/ Category:1980s_films SPARQL • A query Language and ProtocoL for accessing RDF data on the Web SELECT DISTINCT ?x! WHERE {! !?x dcterms:subject ! !<http://dbpedia.org/resource/Category:1980s_horror_films> .! }! SKOS 7 KnowLedge Organizaon Systems (KOS) FoLksonomy ExamPLes #info4302 12 13 What is SKOS? • A modeL for exPressing the basic structure and content of concePt schemes such as thesauri, cLassificaon schemes, taxonomies, folksonomies, and other simiLar tyPes of controLLed vocabuLaries • ALLows concePts to be comPosed and PubLished as Linked Data on the Web • Hides the comPLexity of OWL - easy to use SKOS ConcePts are... • ... idenfied by URIs • ... LabeLed with 1..* natural Language strings • ... documented with various tyPes of notes • ... seman6calLy linked to each other • ... aggregated into concePt schemes ExamPLe SKOS ConcePt skos:ConcePt • ConcePts are – the units of thought: ideas, meanings, categories of objects, etc. – abstract en66es which are indePendent of the terms used to LabeL them skos:Concept rdf:type lcsh:sh2007025344#concept
    [Show full text]
  • The YAGO Knowledge Base
    The YAGO Knowledge Base Fabian M. Suchanek Tel´ ecom´ ParisTech University Paris, France See the original SVG slides 1 A bit of history... Unknown painter 2 2 YAGO: First large KB from Wikipedia Elvis Presley Elvis Presley was one AmericanSinger of the best blah blah blub blah don’t read type this, listen to the speaker! blah blah blah ElvisPresley blubl blah you are still Born: 1935 born bornIn reading this! blah blah In: Tupelo blah blah blabbel blah ... 1935 Tupelo won Categories: locatedIn AcAward Rock&Roll, American Singers, USA Academy Award winners... 3 3 Type Hierarchy Singer from Wikipedia categories, American Singer Male Singer automatically selected Male American Singer Shallow noun Male American Singer of German Origin phrase parsing gender Regular male expressions livesIn USA + demonyms 4 4 Type Hierarchy continued Entity ... from WordNet Organism ... Person by linking Wikipedia & Singer WordNet ... 5 5 Type Hierarchy continued Entity ... ... ... from WordNet Organisation Organism Location ... ... ... University Village Person by linking Wikipedia & Singer WordNet ... In total: 650k classes • every entity has 1 class • ≥ link precision of 97% • 6 6 Knowledge Representation marriedTo Fact Id Subject Predicate Object <fact 42> <Elvis> <marriedTo> <Priscilla> <fact 43> <fact 42> <since> ”1967-05-01”8sd:date <fact 44> <fact 42> <until> ”1973-08-18”8sd:date <fact 45> <fact 42> <extractedFrom> <http://wikipedia...> <fact 46> <fact 45> <byTechnique> ”Infobox “spouse”” (exported to RDF w/out factIds) (Example for illustration only)
    [Show full text]
  • Meet the Data
    Chapter 2 Meet the Data This chapter introduces the basic types of data sources, as well as specific datasets and resources, that we will be working with in later chapters of the book. These may be placed on a spectrum of varying degrees of structure, from unstructured to structured data, as shown in Fig. 2.1. Fig. 2.1 The data spectrum On the unstructured end of the spectrum we have plain text. Typically, these are documents written in natural language.1 As a matter of fact, almost any type of data can be converted into plain text, including web pages, emails, spreadsheets, and database records. Of course, such a conversion would result in an undesired loss of internal document structure and semantics. It is nevertheless always an option to treat data as unstructured, by not making any assumptions about the particular data format. Search in unstructured text is often referred to as full-text search. On the opposite end of the spectrum there is structured data, which is typically stored in relational databases; it is highly organized, tabular, and governed by a strict schema. Search in this type of data is performed using formal query languages, like SQL. These languages allow for a very precise formulation of information needs, but require expert knowledge of the query language and of the underlying database schema. This generally renders them unsuitable for ordinary users. The data we will mostly be dealing with is neither of two extremes and falls somewhere “in the middle.” Therefore, it is termed semi-structured.Itis 1Written in natural language does not imply that the text has to be grammatical (or even sensible).
    [Show full text]
  • YAGO: a Core of Semantic Knowledge Unifying Wordnet and Wikipedia Fabian MS, Gjergji K, Gerhard WE WWW, 2007
    YAGO: A Core of Semantic Knowledge Unifying WordNet and Wikipedia Fabian MS, Gjergji K, Gerhard WE WWW, 2007 Mengdi Huang ([email protected]) 02/11/2019 Problem • The Wikipedia categories’ hierarchy is barely useful for ontological purposes. – eg: Zidane is in the super-category named ”Football in France”, but Zidane is a football player and not a football • Despite its clean taxonomy of concepts, WordNet lacks knowledge about individuals like persons, organizations, products, etc. with their semantic relationships; its number of facts is also limited. • Disadvantages of current Ontologies – Not scalable to Non-Canonical Facts (TextRunner) – Non-defined Relations, and Domains (DBPedia) – Evaluation Results Missing (SemanticWikipedia) 2 Motivation • Given the differences between existing ontologies and their disadvantages, would it help to combine them? • Yes! Applications that utilize ontology can boost their performance if a high coverage and quality ontology with knowledge from several sources is available. • Can we build a new ontology from multiple sources that profit, on one hand, from the vast amount of individuals known to Wikipedia, while exploiting, on the other hand, the clean taxonomy of concepts from WordNet? 3 Motivation • YAGO Facts: – Input: Wikipedia infoboxes, WordNet (and GeoNames in later versions) – Output: KG with 1 million entities and 5 million facts (2007) → KG with 350K entity types, 10M entities, 120M facts (2016) Thomas Rebele, Fabian M. Suchanek et. al, ISWC 2016 – Temporal and spatial information • Challenges – The results of automatic information extraction approach contain many false positives, e.g., IsA(Aachen Cathedral, City), thus requires heavy quality control – Man-made ontologies suffer from low coverage, high cost for assembly and quality assurance, and fast aging.
    [Show full text]
  • Using YAGO for the Humanities
    Using YAGO for the Humanities Thomas Rebele1, Arash Nekoei2, and Fabian M. Suchanek1 1 Telecom ParisTech, France 2 University of Stockholm, Sweden Abstract. In this paper, we study how data from the Semantic Web can be used for case studies in the Humanities. We conduct a proof of concept, using the YAGO knowledge base to study life expectancy, birth rates, and the age at childbirth over time. We also discuss the information extraction methods that we used to make YAGO sufficiently complete for these analyses to work. 1 Introduction Recent advances in the Semantic Web have led to more and more data about the real world in structured form. This data is stored in knowledge bases (KBs). In parallel to this development, the Digital Humanities have established themselves as a research field in their own right, using computational methods to support researchers in their quest to understand history, language, or the arts. Recently, these fields of research have found common ground [1, 9], with the insight that the data of the Semantic Web can help deliver answers to questions in the Digital Humanities. In this paper, we want conduct a proof of concept, and investigate in how far data from YAGO [11], one of the largest general-purpose KBs on the Semantic Web, could help study questions of history and society. Our first observation is that the data in YAGO is very incomplete. For exam- ple, only half of the people in YAGO have a place of residence.3 Therefore, our first challenge is to make the data more complete.
    [Show full text]
  • YAGO2: a Spanally and Temporally Enhanced Knowledge Base from Wikipedia (J. Hoffar
    YAGO2: A spaally and temporally enhanced knowledge base from Wikipedia (J. Hoffart et al, 2012) presented by Gabe Radovsky YAGO2: overview • temporally/spaally anchored knowledge base • built automacally from Wikipedia, GeoNames, and WordNet • contains 447 million facts about 9.8 million en++es • human evaluators judged 97.8% of facts correct The (original) YAGO knowledge base • introduced in 2007 • automacally constructed from Wikipedia – each ar+cle in Wikipedia became an en+ty • about 100 manually defined relaons – e.g. wasBornOnDate, locatedIn, hasPopulaon • used SPO (subject, predicate, object) triples to represent facts – reificaon: every fact given an iden+fier, e.g. wasFoundIn(fact, Wikipedia) YAGO2: mo+vaon • WordNet/other lexical resources: – manually compiled – knows that ‘musician’ is a hyponym of ‘human’; doesn’t know that Leonard Cohen is a musician • Wikipedia/GeoNames – very large collec+ons of (semi-)structured data – advances in informaon extrac+on make them easier to mine YAGO2: mo+vaon • “...current state-of-the-art knowledge bases are mostly blind to the temporal dimension” (29). • e.g. (knowing that Abraham Lincoln was born in 1809 and died in 1865) != (knowing that Abraham Lincoln was alive in 1850) www.wolframalpha.com www.wolframalpha.com YAGO2: contribu+on • top-down ontology “with the goal of integrang en+ty-relaonship-oriented facts with the spaal and temporal dimensions” (29). • new representaon model: SPOTL tuples – (SPO [subject, predicate, object] + +me + locaon) • frameworks for extrac+ng knowledge from structured or unstructured text Extrac+on architecture for YAGO2 • factual rules – “declarave translaons of all the manually defined excep+ons and facts that the previous YAGO code contained” (30) • implicaon rules – e.g.
    [Show full text]
  • Psparql: a Querying Language for Probabilistic RDF (Extended Abstract)
    pSPARQL: A Querying Language for Probabilistic RDF (Extended Abstract) Hong Fang1 and Xiaowang Zhang2;3;4;? 1 College of Arts and Sciences, Shanghai Polytechnic University, Shanghai 201209, China 2 School of Computer Science and Technology, Tianjin University, Tianjin 300350, China 3 Tianjin Key Laboratory of Cognitive Computing and Application, Tianjin 300350, China 4 Key Laboratory of Computer Network and Information Integration (Southeast University), Ministry of Education, Nanjing 211189, China Abstract. In this paper, we present a querying language for probabilistic RDF databases, where each triple has a probability, called pSRARQL, built on SPAR- QL, recommended by W3C as a querying language for RDF databases. First- ly, we present the syntax and semantics of pSPARQL. Secondly, we define the query problem of pSPARQL corresponding to probabilities of solutions. Finally, we show that the query evaluation of general pSPARQL patterns is PSPACE- complete. 1 Introduction Resource Description Framework (RDF)5 is the standard data model in the Se- mantic Web. In our real world, RDF data possibly contains some uncertainty data due to the diversity of data sources such as YAGO6. For instance, some RDF data is generated from raw data via knowledge extraction and machine learning. However, RDF model itself provides little support for uncertain data [3]. There are some approaches to querying over probabilistic RDF [1,2,4,3]. Though those approaches can query probabilistic RDF, they cannot support S- PARQL 7 which, recommended by W3C, has become the standard language for querying RDF data since 2008. Indeed, SPARQL has been applied to query probabilistic ontologies [6]. In this paper, we present an extended querying language (called pSPARQL: probabilistic SPARQL) for probabilistic RDF databases with supporting SPAR- QL.
    [Show full text]