Survey of Tools for Linked Data Consumption

Total Page:16

File Type:pdf, Size:1020Kb

Survey of Tools for Linked Data Consumption Semantic Web 0 (2017) 1–42 1 IOS Press Survey of Tools for Linked Data Consumption Jakub Klímek a;∗ Petr Škoda a Martin Necaskýˇ a a Faculty of Mathematics and Physics, Charles University in Prague Malostranské námˇestí25, 118 00 Praha 1, Czech Republic E-mail: [email protected], [email protected], [email protected] Abstract. There is lots of data published as Linked (Open) Data (LOD/LD). At the same time, there is also a multitude of tools for publication of LD. However, potential LD consumers still have difficulty discovering, accessing and exploiting LD. This is because compared to consumption of traditional data formats such as XML and CSV files, there is a distinct lack of tools for consumption of LD. The promoters of LD use the well-known 5-star Open Data deployment scheme to suggest that consumption of LD is a better experience once the consumer knows RDF and related technologies. This suggestion, however, falls short when the consumers search for an appropriate tooling support for LD consumption. In this paper we define a LD consumption process. Based on this process and current literature, we define a set of 36 requirements a hypothetical Linked Data Consumption Platform (LDCP) should ideally fulfill. We cover those requirements with a set of 93 evaluation criteria. We survey 110 tools identified as potential candidates for LDCP, eliminating them in 4 rounds until 9 candidates for LDCP remain. We evaluate the 9 candidates using our 93 criteria. Based on this evaluation we show which parts of the LD consumption process are covered by the 9 candidates. We also show that there are important LD consumption steps which are not sufficiently covered by existing tools. The authors of LDCP implementations may use our paper to decide about directions of future development of their tools. The paper can also be used as an introductory text to LD consumption. Keywords: Linked Data, RDF, dataset, consumption, discovery, visualization, tool, survey 1. Introduction that the non-trivial effort put into publishing the data as LD will bring benefits mainly to them, the users of A considerable amount of data represented in the the data. These benefits, compared to 3-star Open Data, form of Linked (Open) Data (LOD/LD) is now avail- include better described and understandable data and able on the Web, or in enterprise knowledge graphs. its schema, safer data integration thanks to the global More and more data publishers are convinced that the character of the IRIs and shared vocabularies, better additional effort put into the last 2 stars of the now reuse of tools, more context thanks to the links and the widely accepted and promoted 5-star Open Data deploy- ability to (re)use only parts of the data. ment scheme1 will bring the promised benefits to the To encourage Open Data users to consume LD there users of their data. The publishers rightfully expect that is a need for a platform which will help them with such the users will appreciate the 5-star data much more than, consumption and which will exploit all the benefits of e.g., the 3-star CSV files files given that the proper pub- LD. It is not so important whether such platform is a lication of 5-star data is considerably harder and more single software tool or a set of compatible and reusable costly to achieve. Naturally, the expectations of users software tools which can be combined as needed. The of LD are quite high as the promoters of LD promise necessity of building such a platform is also identified in [107], where the authors try to build a toolset us- *Corresponding author. E-mail: [email protected]. able by users to access LD. In this paper we call such 1http://5stardata.info/ a hypothetical platform a Linked Data Consumption 1570-0844/17/$35.00 c 2017 – IOS Press and the authors. All rights reserved Platform (LDCP). In the broadest sense, LDCP should using the whole set of criteria. The evaluation is not a support users in discovering, loading and processing user based study and, therefore, it does not evaluate the LD datasets with the aim of either getting some useful usability of the tools. visual output, which can be presented to lay users, or This paper is based on our introductory study of getting a machine-readable output, which can then be the problem [72]. Since then, we have better defined used in many existing non-RDF processing tools avail- requirements, better defined survey methodology and able today for data analysis and visualization. In the we survey considerably more tools which are also more minimal sense, LDCP should be able to load a given LD recent, which we present here. dataset or its part and provide some non-RDF output, This paper is structured as follows. In Section 2 we either a visualization or a machine-readable output in a provide a motivating example of a user scenario in non-RDF format, such as a CSV file. which a journalist wants to consume LD and expects The aim of this paper is to identify existing usable a platform that will enable him to enjoy the promised software tools which could serve as LDCP and assess LD benefits. In Section 3 we describe our evaluation the extent of support they could provide to the users methodology. In Section 4 we define the LD consump- regarding LD consumption. tion process, describe the identified requirements and In the first part of the paper, we describe the LD con- cover them with evaluation criteria. In Section 5 we sumption process and we identify requirements which identify existing tools and evaluate them using the eval- specify in detail how individual activities in the LD con- uation criteria. In Section 6 we survey related studies sumption process should be supported by LDCP. The of LD consumption and in Section 7 we conclude. requirements are identified based on a motivating sce- nario, existing W3C Recommendations and research ar- eas and approaches published in recent scientific litera- 2. Motivating Example ture. This paper is not a detailed and complete survey of existing research literature related to LD consumption. To motivate our work, let us suppose a data journalist We use the existing literature to support the proposed who is used to working with 3-star Open Data, i.e. non- requirements. In other words, the individual proposed RDF Open Data available on the web such as HTML requirements are not novel. Each requirement is either pages, CSV tables, etc. He needs to collect data about back by an existing W3C Recommendation and/or the population of cities in Europe and display an informa- possibility of fulfilling the requirement is described in tive map in an article he is working on. The intended the literature. We compose the requirements into the audience of the article is also statisticians who are used LD consumption process. Then, for each requirement to CSVs, so the journalist also wants to publish the we define a set of criteria to be used to decide whether underlying data as a CSV file attached to the article. a given software tool fulfills the requirement or not. In He is used to working with 3-star open data, which total, the paper presents 36 requirements covered by 93 means that he has to know where to find the necessary evaluation criteria. data sources, e.g. population statistics and descriptive In the second part of the paper, we identify existing data about cities including their location, he has to usable software tools which we classify as LDCP candi- download them, integrate them and process them to get dates. We did a systematic review to identify these soft- the required output, i.e. a visualization and a CSV file. ware tools. Then we filtered these software tools only The journalist now wants to use LD because he heard to those which fulfill the following conditions which that it is the highest possible level of openness accord- define a minimalistic LDCP described above: ing to the 5-star Open Data deployment scheme. He ex- pects the experience of working with LD will be some- – It is possible to install the tool locally or its public how better than the experience of working with 3-star instance is available on the web. data once he learns RDF and related technologies. Let – It is possible to load RDF data into the tool. – It is possible to get some output from the tool us now suppose that he has learned them. Now he ex- which can be used by a non-RDF expert, such pects to get better access to data and better experience as a visualization or an output in a 3-star data working with the data. representation Naturally, the user expects a tool that can be used to work with such data and which will bring the promised As a result we identify 9 usable software tools which benefits of LD. Ideally, the tool would be an integrated we classify as minimal LDCP and evaluate them further platform that supports the whole process of LD con- sumption starting from identification of data sources, Let us now suppose that the user has all the datasets ending with the resulting visualization or processed needed for his goal. There could be an interoperabil- data ready to be used by other non-RDF tools while ity issue caused by the existence of multiple vocabu- using the benefits of LD where applicable. laries describing the same domain. In our example, it Intuitively, the user needs to find data about cities in could be that the dataset containing locations of the Europe, i.e.
Recommended publications
  • Applications
    Applications CSE 595 – Semantic Web Instructor: Dr. Paul Fodor Stony Brook University http://www3.cs.stonybrook.edu/~pfodor/courses/cse595.html GoodRelations BBC Artists BBC World Cup Website Lecture Outline Government Data New York Times Sig.ma and Sindice Swoogle OpenCalais Schema.org data.world Elsevier Audi Data Integration Swiss Life EnerSearch E-Learning Web Services Multimedia Collection Indexing at Scotland Yard Online Procurement at Daimler Device Interoperability at Nokia 2 Publication Management @ Semantic Web Primer GoodRelations E-commerce, and in particular Business-to-Consumer (B2C) e- commerce, has been one of the main drivers behind the rapid adoption of the World Wide Web in everyday live It is now commonplace to see URLs listed on storefronts and goods vehicles Taking the UK as an example, the B2C market has grown from £87 million in April 2000 to £68.4 billion by the end of 2009, a thousand-fold increase over a single decade USA 2017 B2C market was $660 billion, but the growth is decreasing 3 statista.com @ Semantic Web Primer GoodRelations E-commerce marketplace is suffering from all the deficits of the traditional web: E-commerce websites are typically generated from structured information systems, listing price, availability, type of product, delivery options, etc., but by the time this information reaches the company’s web pages, it has been turned into HTML and all machine-interpretable structure has disappeared, with the result that machines can no longer distinguish a price from a product-code
    [Show full text]
  • V a Lida T in G R D F Da
    Series ISSN: 2160-4711 LABRA GAYO • ET AL GAYO LABRA Series Editors: Ying Ding, Indiana University Paul Groth, Elsevier Labs Validating RDF Data Jose Emilio Labra Gayo, University of Oviedo Eric Prud’hommeaux, W3C/MIT and Micelio Iovka Boneva, University of Lille Dimitris Kontokostas, University of Leipzig VALIDATING RDF DATA This book describes two technologies for RDF validation: Shape Expressions (ShEx) and Shapes Constraint Language (SHACL), the rationales for their designs, a comparison of the two, and some example applications. RDF and Linked Data have broad applicability across many fields, from aircraft manufacturing to zoology. Requirements for detecting bad data differ across communities, fields, and tasks, but nearly all involve some form of data validation. This book introduces data validation and describes its practical use in day-to-day data exchange. The Semantic Web offers a bold, new take on how to organize, distribute, index, and share data. Using Web addresses (URIs) as identifiers for data elements enables the construction of distributed databases on a global scale. Like the Web, the Semantic Web is heralded as an information revolution, and also like the Web, it is encumbered by data quality issues. The quality of Semantic Web data is compromised by the lack of resources for data curation, for maintenance, and for developing globally applicable data models. At the enterprise scale, these problems have conventional solutions. Master data management provides an enterprise-wide vocabulary, while constraint languages capture and enforce data structures. Filling a need long recognized by Semantic Web users, shapes languages provide models and vocabularies for expressing such structural constraints.
    [Show full text]
  • A Field Guide to Web Apis by Kin Lane Contents
    A field guide to web APIs By Kin Lane Contents Executive summary 4 What are APIs used for? 5 Open data . 5 Websites . 5 Mobile . 5. Automobiles . 6. Homes and buildings . 6 Why are web APIs different? 7 They build on existing web architecture . 7. Intuitive resources . 7 Simplicity rules . 8. Easy to understand for developers and even nondevelopers . 8. Self-service resources . 8 . History of web APIs 9 Commerce . 9 . Social . 9 . Cloud computing . .9 . Mobile . .10 . What technology goes into an API? 11 REST . 11. JSON . 11 Security . 11 . Keys . 11 . Basic auth . 12 Open authorization . 12 . Webhooks . 12 Deploying your web API 13 Do-it-yourself approaches . 13 Cloud solutions . 13 . Enterprise gateways . 13 . Established practices for managing APIs 14 Self-service . 14 . Getting started . .14 . Documentation . 15 . Code samples . 15. 2 A field guide to web APIs Support and feedback loops . 15 . The legal aspect . 15. Developer dashboard . 16 Marketing and API evangelism 17 Goals . 17 User engagement . .17 . Blogging . 17 Landscape analysis . 18 . GitHub . .18 . Social . 18. Events . 19. The future of web APIs 20 API aggregation . 20 . Real-time APIs . 20. Backend as a Service (BaaS) . 20 . Automation . 20 Voice . 21. Internet of things . 21. Cloud trends 22 Maturity of IaaS layer . 22. Opportunities in the PaaS layer . .22 . Key takeaways 23 About Kin Lane 23 3 A field guide to web APIs Executive summary A new breed of web API has emerged, delivering a vision of a lightweight, low-cost approach to connect devices and allowing applications to exchange data efficiently. This research report is a field guide for web API providers, developers, and even nondevelopers .
    [Show full text]
  • Semantics Developer's Guide
    MarkLogic Server Semantic Graph Developer’s Guide 2 MarkLogic 10 May, 2019 Last Revised: 10.0-8, October, 2021 Copyright © 2021 MarkLogic Corporation. All rights reserved. MarkLogic Server MarkLogic 10—May, 2019 Semantic Graph Developer’s Guide—Page 2 MarkLogic Server Table of Contents Table of Contents Semantic Graph Developer’s Guide 1.0 Introduction to Semantic Graphs in MarkLogic ..........................................11 1.1 Terminology ..........................................................................................................12 1.2 Linked Open Data .................................................................................................13 1.3 RDF Implementation in MarkLogic .....................................................................14 1.3.1 Using RDF in MarkLogic .........................................................................15 1.3.1.1 Storing RDF Triples in MarkLogic ...........................................17 1.3.1.2 Querying Triples .......................................................................18 1.3.2 RDF Data Model .......................................................................................20 1.3.3 Blank Node Identifiers ..............................................................................21 1.3.4 RDF Datatypes ..........................................................................................21 1.3.5 IRIs and Prefixes .......................................................................................22 1.3.5.1 IRIs ............................................................................................22
    [Show full text]
  • Product Schema, SEO, Structured Data | Caliber Media Group & Schema.Org Products | Product Schema | SEO
    ProductCamp: Product Schema, SEO, Structured Data | Caliber Media Group & Schema.org Products | Product Schema | SEO Caliber Media Group presented at Product Camp Southern California 2014, and Caliber also volunteered its services towards this conference’s success. Caliber Media Group ProductCamp: Product Schema, SEO, Structured Data | Caliber Media Group & Schema.org Products | Schema | Examples Resources | Tools | Readings 2 ProductCamp SoCal 2014 & Schema: Google, after Schema ProductCamp SoCal 2014 & Schema: Google via Disconnect & Yahoo, after Schema , ProductCamp SoCal 2014 & Schema: How did Caliber help ProductCamp beat all of the other events? ProductCamp SoCal 2014 & Schema: So what was inside the pages? ProductCamp SoCal 2014 & Schema: So what was inside the pages? ProductCamp SoCal 2014 & Schema: So what was inside the pages? Search => Schema, with JSON-LD Search & Schema: When & Who? Very short version… Martin Hepp http://www.heppnetz.de/projects/goodrelations/ ProductCamp SoCal 2014 & Schema: When & Who? Short version…here they come… ProductCamp SoCal 2014 & Schema: GoodRelations added… Search & Schema: Why, and… What does Schema do for me? http://www.schema.org ProductCamp SoCal 2014 & Schema: Product since this is ProductCamp http://schema.org/Product ProductCamp SoCal 2014 & Schema: An actual Product, since this is ProductCamp… ProductCamp SoCal 2014 & Schema: Recall, Products as Structured Data Structure -- Hierarchy -- Structured Data Search & Schema: Show me Schema examples… Schema: Product ontology | microdata | Knowledge Graph ProductCamp SoCal 2014 & Schema: Product ProductCamp SoCal 2014 & Schema: Google ProductCamp SoCal 2014 & Schema: Product Schema locally ProductCamp SoCal 2014 & Schema: Product locally ProductCamp SoCal 2014 & Schema: Local Schema gone wild… Search & Product Schema | Knowledge Graph from Google Knowledge Vault – where? Search & Product Schema | Knowledge Graph – where? Google Tables.
    [Show full text]
  • Seamless Interoperability and Data Portability in the Social Web for Facilitating an Open and Heterogeneous Online Social Network Federation
    Seamless Interoperability and Data Portability in the Social Web for Facilitating an Open and Heterogeneous Online Social Network Federation vorgelegt von Dipl.-Inform. Sebastian Jürg Göndör geb. in Duisburg von der Fakultät IV – Elektrotechnik und Informatik der Technischen Universität Berlin zur Erlangung des akademischen Grades Doktor der Ingenieurwissenschaften - Dr.-Ing. - genehmigte Dissertation Promotionsausschuss: Vorsitzender: Prof. Dr. Thomas Magedanz Gutachter: Prof. Dr. Axel Küpper Gutachter: Prof. Dr. Ulrik Schroeder Gutachter: Prof. Dr. Maurizio Marchese Tag der wissenschaftlichen Aussprache: 6. Juni 2018 Berlin 2018 iii A Bill of Rights for Users of the Social Web Authored by Joseph Smarr, Marc Canter, Robert Scoble, and Michael Arrington1 September 4, 2007 Preamble: There are already many who support the ideas laid out in this Bill of Rights, but we are actively seeking to grow the roster of those publicly backing the principles and approaches it outlines. That said, this Bill of Rights is not a document “carved in stone” (or written on paper). It is a blog post, and it is intended to spur conversation and debate, which will naturally lead to tweaks of the language. So, let’s get the dialogue going and get as many of the major stakeholders on board as we can! A Bill of Rights for Users of the Social Web We publicly assert that all users of the social web are entitled to certain fundamental rights, specifically: Ownership of their own personal information, including: • their own profile data • the list of people they are connected to • the activity stream of content they create; • Control of whether and how such personal information is shared with others; and • Freedom to grant persistent access to their personal information to trusted external sites.
    [Show full text]
  • Navigating the New Streaming API Landscape
    WHITE PAPER Navigating the new streaming API landscape 10 ways real-time, event-driven architectures are redefining the future of business Navigating the new streaming API landscape Streaming: it’s much Streaming APIs that deliver data to web, mobile, and device applications more than video have been evolving for many years, but today an important convergence is happening between regular web APIs and streaming technology. The result When you mention streaming to is the advent of the event-driven architecture, and it’s redefining the future anyone in the context of the web, of business. they immediately think of Netflix and YouTube. This paper is not There are a lot of online services, tools, and devices that use the word about the streaming of video “streaming,” making for a crazy mix of information that will make even the content. It’s about widening your most seasoned technologist’s head spin. The objective of this paper is to perspective on what real-time help anyone — technical or not — navigate the fast-growing and ever-evolving streaming can be, and the many streaming API landscape. We’ll walk you through some of the most common reasons a company, use cases for making data available in a real-time stream to help you organization, institution, or understand how and when it makes sense to invest in your streaming government agency would want API infrastructure. to invest in streaming APIs and event-driven architecture. 10 ways real-time, event-driven architectures are redefining the future of business To help paint a picture of why streaming data using APIs isn’t just about video, financial data, or news, we’ll take a look at 10 areas of significant streaming technology adoption.
    [Show full text]
  • Bibliography of Erik Wilde
    dretbiblio dretbiblio Erik Wilde's Bibliography References [1] AFIPS Fall Joint Computer Conference, San Francisco, California, December 1968. [2] Seventeenth IEEE Conference on Computer Communication Networks, Washington, D.C., 1978. [3] ACM SIGACT-SIGMOD Symposium on Principles of Database Systems, Los Angeles, Cal- ifornia, March 1982. ACM Press. [4] First Conference on Computer-Supported Cooperative Work, 1986. [5] 1987 ACM Conference on Hypertext, Chapel Hill, North Carolina, November 1987. ACM Press. [6] 18th IEEE International Symposium on Fault-Tolerant Computing, Tokyo, Japan, 1988. IEEE Computer Society Press. [7] Conference on Computer-Supported Cooperative Work, Portland, Oregon, 1988. ACM Press. [8] Conference on Office Information Systems, Palo Alto, California, March 1988. [9] 1989 ACM Conference on Hypertext, Pittsburgh, Pennsylvania, November 1989. ACM Press. [10] UNIX | The Legend Evolves. Summer 1990 UKUUG Conference, Buntingford, UK, 1990. UKUUG. [11] Fourth ACM Symposium on User Interface Software and Technology, Hilton Head, South Carolina, November 1991. [12] GLOBECOM'91 Conference, Phoenix, Arizona, 1991. IEEE Computer Society Press. [13] IEEE INFOCOM '91 Conference on Computer Communications, Bal Harbour, Florida, 1991. IEEE Computer Society Press. [14] IEEE International Conference on Communications, Denver, Colorado, June 1991. [15] International Workshop on CSCW, Berlin, Germany, April 1991. [16] Third ACM Conference on Hypertext, San Antonio, Texas, December 1991. ACM Press. [17] 11th Symposium on Reliable Distributed Systems, Houston, Texas, 1992. IEEE Computer Society Press. [18] 3rd Joint European Networking Conference, Innsbruck, Austria, May 1992. [19] Fourth ACM Conference on Hypertext, Milano, Italy, November 1992. ACM Press. [20] GLOBECOM'92 Conference, Orlando, Florida, December 1992. IEEE Computer Society Press. http://github.com/dret/biblio (August 29, 2018) 1 dretbiblio [21] IEEE INFOCOM '92 Conference on Computer Communications, Florence, Italy, 1992.
    [Show full text]
  • The Application of Semantic Web Technologies to Content Analysis in Sociology
    THEAPPLICATIONOFSEMANTICWEBTECHNOLOGIESTO CONTENTANALYSISINSOCIOLOGY MASTER THESIS tabea tietz Matrikelnummer: 749153 Faculty of Economics and Social Science University of Potsdam Erstgutachter: Alexander Knoth, M.A. Zweitgutachter: Prof. Dr. rer. nat. Harald Sack Potsdam, August 2018 Tabea Tietz: The Application of Semantic Web Technologies to Content Analysis in Soci- ology, , © August 2018 ABSTRACT In sociology, texts are understood as social phenomena and provide means to an- alyze social reality. Throughout the years, a broad range of techniques evolved to perform such analysis, qualitative and quantitative approaches as well as com- pletely manual analyses and computer-assisted methods. The development of the World Wide Web and social media as well as technical developments like optical character recognition and automated speech recognition contributed to the enor- mous increase of text available for analysis. This also led sociologists to rely more on computer-assisted approaches for their text analysis and included statistical Natural Language Processing (NLP) techniques. A variety of techniques, tools and use cases developed, which lack an overall uniform way of standardizing these approaches. Furthermore, this problem is coupled with a lack of standards for reporting studies with regards to text analysis in sociology. Semantic Web and Linked Data provide a variety of standards to represent information and knowl- edge. Numerous applications make use of these standards, including possibilities to publish data and to perform Named Entity Linking, a specific branch of NLP. This thesis attempts to discuss the question to which extend the standards and tools provided by the Semantic Web and Linked Data community may support computer-assisted text analysis in sociology. First, these said tools and standards will be briefly introduced and then applied to the use case of constitutional texts of the Netherlands from 1884 to 2016.
    [Show full text]
  • Dezvoltarea Aplicațiilor Web La Nivel De Client
    inginerie inginerie Dezvoltarea Dezvoltarea aplicațiilor Web software în contextulsoftware la niveldeclient JavaScript Dr. Sabin Buragaprofs.info.uaic.ro/~busaco busaco ~ “There are two ways to write error-free programs; / only the third one works.” profs.info.uaic.ro Alan J. Perlis Dr. Sabin Sabin Buraga Dr. busaco ~ / Ce instrumente software și biblioteci JS pot fi folosite în contextul programării Web profs.info.uaic.ro la nivel de client? Dr. Sabin Sabin Buraga Dr. instrumente busaco ~ / Editare de cod și dezvoltare de aplicații Web Depanare profs.info.uaic.ro Testare Documentare a codului Compresie Optimizare Sabin Buraga Dr. instrumente: editare busaco ~ / Editoare + medii integrate (IDE) pentru desktop: Atom Editor, Brackets, JS Development Tools (Eclipse), Sublime Text, Visual Studio Code, WebStorm etc. profs.info.uaic.ro Disponibile pe Web – în cloud: Cloud9 IDE, Codepen, Codio, JS Bin, JS Fiddle, Plunker,… Dr. Sabin Sabin Buraga Dr. unele oferă și partajarea codului-sursă cu alți dezvoltatori a se considera și github.com/sorrycc/awesome-javascript#editors instrumente: depanare busaco ~ / Utilizarea consolei JS obiectul console oferit de browser profs.info.uaic.ro console.spec.whatwg.org (Living Standard, 25 oct. 2018) metode utile: log (), error (), warn (), info (), assert (), time (), timeEnd (), trace (), group (), groupEnd (), table (), dir (), dirxml () Sabin Buraga Dr. developer.mozilla.org/Web/API/Console instrumente: depanare busaco ~ / profs.info.uaic.ro de studiat și M. Burgess, Beyond console.log(), nov. 2018 medium.com/@mattburgess/beyond-console-log-2400fdf4a9d8 Dr. Sabin Sabin Buraga Dr. instrumente: depanare busaco ~ / Instrumentele de depanare profs.info.uaic.ro incluse în navigatoarele Web moderne: developers.google.com/web/tools/chrome-devtools/ developer.mozilla.org/docs/Tools docs.microsoft.com/en-us/microsoft-edge/devtools-guide developer.apple.com/safari/tools/ Dr.
    [Show full text]
  • 15 Years of Semantic Web: an Incomplete Survey
    Ku¨nstl Intell (2016) 30:117–130 DOI 10.1007/s13218-016-0424-1 TECHNICAL CONTRIBUTION 15 Years of Semantic Web: An Incomplete Survey 1 2 Birte Glimm • Heiner Stuckenschmidt Received: 7 November 2015 / Accepted: 5 January 2016 / Published online: 23 January 2016 Ó Springer-Verlag Berlin Heidelberg 2016 Abstract It has been 15 years since the first publications today’s Web: Uniform Resource Identifiers (URIs) for proposed the use of ontologies as a basis for defining assigning unique identifiers to resources, the HyperText information semantics on the Web starting what today is Markup Language (HTML) for specifying the formatting known as the Semantic Web Research Community. This of Web pages, and the Hypertext Transfer Protocol (HTTP) work undoubtedly had a significant influence on AI as a that allows for the retrieval of linked resources from across field and in particular the knowledge representation and the Web. Typically, the markup of standard Web pages Reasoning Community that quickly identified new chal- describes only the formatting and, hence, Web pages and lenges and opportunities in using Description Logics in a the navigation between them using hyperlinks is targeted practical setting. In this survey article, we will try to give towards human users (cf. Fig. 1). an overview of the developments the field has gone through In 2001, Tim Berners-Lee, James Hendler, and Ora in these 15 years. We will look at three different aspects: Lassila describe their vision for a Semantic Web [5]: the evolution of Semantic Web Language Standards, the The Semantic Web is not a separate Web but an evolution of central topics in the Semantic Web Commu- extension of the current one, in which information is nity and the evolution of the research methodology.
    [Show full text]
  • Designing an Ontology for Managing the Diets of Hypertensive Individuals
    Designing an Ontology for Managing the Diets of Hypertensive Individuals A thesis submitted to the College of Communication and Information of Kent State University in partial fulfillment of the requirements for the Master of Library and Information Science and Master of Science dual degree program By Julaine Clunis January, 2016 Thesis Written By Julaine Clunis BSc., Northern Caribbean University, 2005 M.L.I.S., Kent State University, 2016 M.S., Kent State University, 2016 Approved By ___________________________________________ Marcia Lei Zeng, Advisor ___________________________________________ Jeffrey W. Fruit, Director, School of Library and Information Science ___________________________________________ Amy Reynolds, Dean, College of Communication and Information Table of Contents List of Figures ................................................................................................................................ v List of Tables ............................................................................................................................... vii List of Acronyms and Abbreviations ....................................................................................... viii Glossary of Terms ......................................................................................................................... x Terms Relating to Health and Medicine ............................................................................................. xii Acknowledgements ...................................................................................................................
    [Show full text]