Semantic Web Methods for Knowledge Management

Total Page:16

File Type:pdf, Size:1020Kb

Semantic Web Methods for Knowledge Management Semantic Web Methods for Knowledge Management Zur Erlangung des akademischen Grades eines Doktors der Wirtschaftswissenschaften (Dr. rer. pol.) an der Fakultät für Wirtschaftswissenschaften der Universität Fridericiana zu Karlsruhe (TH) genehmigte DISSERTATION von Diplom-Informatiker Stefan Decker Tag der mündlichen Prüfung: 22. Februar 2002 Referent: Prof. Dr. Rudi Studer 1. Korreferent: Prof. Dr. Peter Knauth 2. Korreferent: Prof. Dr. Gio Wiederhold 2002 Karlsruhe What information consumes is rather obvious: It consumes the attention of its recipients. Hence a wealth of information creates a poverty of attention, and a need to allocate that attention efficiently among the overabundance of information sour- ces that might consume it. Nobel Laureate Economist Herbert A. Simon. Acknowledgement The main results of this thesis were obtained during my time at the Institute AIFB at the University of Karlsruhe, Germany. I would like to express my gratitude to my advisor, Prof. Dr. Rudi Studer, for his support, patience, and encouragement throughout my graduate studies. It is not often that one finds an advisor and colleague who always finds the time for listening to the little problems and roadblocks that unavoidably crop up in the course of performing research. I am also grateful to Prof. Dr. Peter Knauth and Prof. Gio Wiederhold who were willing to serve on my dissertation committee on a very short notice, and to Prof. Dr. Wolfried Stucky and Prof. Dr. Andreas Geyer-Schulz, who served on the examination committee. My thanks also go to all the colleagues at the Institute AIFB - in particular to Dr. Michael Erdmann - together we worked on the nitty-gritty details of the Ontobroker architecture, and Michael helped me to shape my ideas about the future of the Web. The vision, energy and dedication of Dr. Dieter Fensel has been invaluable for this thesis and the Ontobroker project - without him that project had not happened. Michael Schenk, Ulf Steinberg, and Thorsten Wolf, who wrote their diploma theses under my supervision, contributed heavily to Ontobroker. Dr. Wolfgang Weitz was always there for a discussion when I got stuck. I'm grateful to Dr. Jürgen Angele for the collaboration during my time in Karlsruhe. I'm especially indebted to Prof. Dr. Gio Wiederhold, who gave me the opportunity to work at Stan- ford University and taught me a lot about research and life in general. I thank Jade Reidy and Michael Sintek for proofreading drafts of this thesis - any remaining errors are of course mine. My parents, Maria and Bernhard, receive my deepest gratitude and love for their dedication and the many years of support during my undergraduate studies that provided the foundation for this work. Last, but not least, I would like to thank my wife Birgit and my children Jana and Robin for their understanding and love during the past years. Their support and encouragement was in the end what made this dissertation possible. v Table of Contents Chapter 1 Introduction . 1 1.1 Motivation . 1 1.1.1 Knowledge Management . 1 1.1.2 The Web . 2 1.2 Goals . 2 1.3 Approaches . 3 1.4 Outline . 4 PART I KNOWLEDGE MANAGEMENT AND ORGANIZATIONAL MEMORIES Chapter 2 Preliminaries: What is Knowledge and Knowledge Management? . 11 2.1 What is Knowledge Management? . 11 2.2 Knowledge in Computer Science . 13 2.2.1 Knowledge in Artificial Intelligence . 13 2.2.2 Ontologies . 15 2.2.3 Representing Procedural Knowledge . 21 2.2.4 Knowledge and Computer Supported Cooperative Work . 22 Chapter 3 Dimensions of Knowledge Management . 25 3.1 Elements of Knowledge Management . 25 3.1.1 Corporate Culture . 25 3.1.2 Organization . 26 3.1.3 People . 28 3.1.4 Technology . 29 3.2 Elements of Knowledge Management II . 29 3.2.1 Defining Knowledge Goals . 29 3.2.2 Knowledge Identification . 29 3.2.3 Knowledge Acquisition . 30 3.2.4 Knowledge Development . 32 3.2.5 Knowledge Sharing and Dissemination . 33 3.2.6 Knowledge Preservation . 34 3.2.7 Knowledge Utilization . 34 3.2.8 Knowledge Assessment . 35 3.3 A Synthesis: The Knowledge Management Cube . 35 Chapter 4 Organizational Memory Information Systems . 37 4.1 IT-Support for Knowledge Management . 37 4.2 Crucial Points for Realizing OMIS . 38 4.2.1 Knowledge Acquisition and Maintenance . 39 4.2.2 Knowledge Integration . 41 4.2.3 Knowledge Retrieval . 42 4.2.4 Differences between OMIS and other Information Systems . 43 4.2.5 OMIS and Expert Systems . 44 vi Chapter 5 A Modeling Schema for Integrating Enterprise Processes and OMIS Support . 45 5.1 Motivation . 45 5.2 CommonKADS . 46 5.2.1 CommonKADS Organization Model . 47 5.2.2 CommonKADS Task Model . 49 5.2.3 CommonKADS Expertise Model . 49 5.3 Relationship between Process Meta Models and Models for Knowledge Based Systems . 51 5.4 Deriving the Schema of OMMM . 53 5.5 Views of OMMM . 54 5.6 Related Work in Enterprise Modeling . 59 5.7 An Application: ERBUS . 62 5.7.1 Introduction . 62 5.7.2 Modeling the Design Process of Industrial Designer . 63 5.7.3 Identifying Tasks to Support . 64 5.7.4 Experiences and Assessment . 65 5.8 Summary . 66 PART II KNOWLEDGE MANAGEMENT WITH ONTOBROKER Chapter 6 Introduction to Ontobroker . 71 6.1 Motivation . 71 6.2 The Architecture of Ontobroker . 74 6.2.1 The Approach . 75 6.2.2 The System . 76 Chapter 7 SiLRI: The Representation Language and System . 79 7.1 Introduction . 79 7.1.1 Higher-Order Logics . 80 7.1.2 Full First-Order Logic-Based Inference Engines . 81 7.1.3 Description Logic . 81 7.1.4 Logic Programming and Deductive Databases . 81 7.2 F-Logic in SiLRI . 82 7.2.1 Syntax of F-Logic in SiLRI . 83 7.3 Semantics of F-Logic . 84 7.3.1 Constructing Domain-Specific Logic Languages . 86 7.3.1.1 Frame-Logic . 88 7.3.1.2 Chronolog . 90 7.3.1.3 C-F-Logic: An Amalgamation of F-logic and Chronolog . 91 7.3.1.4 Assessment of the Approach . 93 7.3.2 Optimized Lloyd-Topor Transformation . 94 7.4 Selecting an Evaluation Procedure and Semantics . 96 7.5 Implementation of SiLRI . 96 7.6 Critical Evaluation of F-Logic as a Representation and Query Language for the Web . 97 vii Chapter 8 Document Annotation . 99 8.1 Metadata . 99 8.2 HTML-A . ..
Recommended publications
  • What Is Your Software Worth? © Gio Wiederhold Stanford University April 2007 a Shorter Version of This Paper Was Published in the Comm
    What is Your Software Worth? © Gio Wiederhold Stanford University April 2007 A shorter version of this paper was published in the Comm. of the ACM, Vol. 49 No.9, Sep. 2006 [Wiederhold:06]. Abstract This article presents a method for valuing software, based on the income that use of that software is expected to generate in the future. It applies well known principles of intellectual property (IP) valuation, sales expectations, software maintenance, product growth, discounting to present value, and the like, always focusing on the specific issues that arise when the benefits of software are to be analyzed. An issue, not dealt with in the literature of valuing intangibles, is that software is continually upgraded. Applying depreciation schedules is the simple solution, but depreciation is taken by purchasers, and does not represent the actual diminution of the inherent IP of software at the supplier. A novel approach, which considers ongoing maintenance and its effects, is presented here. All steps of the process are presented and then integrated via a simple quantitative example. Having a quantitative model on a spreadsheet allows exploration of alternatives. As an example we evaluate a service business model alternative. Some conclusions are drawn that reflect on academic and business practice. 1. Introduction. There exists a voluminous literature on estimation of the cost of producing software, but that literature largely ignores the benefits of using that software [Boehm:81, 00]. Even software engineering management approaches termed `Earned Value Management' only deal with expenses within a development schedule [Abba:97]. While we, as software creators, believe that what we produce is valuable, we are rarely called upon to quantify its benefits [GarmusH:01].
    [Show full text]
  • How to Keep a Knowledge Base Synchronized with Its Encyclopedia Source
    Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence (IJCAI-17) How to Keep a Knowledge Base Synchronized with Its Encyclopedia Source Jiaqing Liang12, Sheng Zhang1, Yanghua Xiao134∗ 1School of Computer Science, Shanghai Key Laboratory of Data Science Fudan University, Shanghai, China 2Shuyan Technology, Shanghai, China 3Shanghai Internet Big Data Engineering Technology Research Center, China 4Xiaoi Research, Shanghai, China [email protected], fshengzhang16,[email protected] Abstract However, most of these knowledge bases tend to be out- dated, which limits their utility. For example, in many knowl- Knowledge bases are playing an increasingly im- edge bases, Donald Trump is only a business man even af- portant role in many real-world applications. How- ter the inauguration. Obviously, it is important to let ma- ever, most of these knowledge bases tend to be chines know that Donald Trump is the president of the United outdated, which limits the utility of these knowl- States so that they can understand that the topic of an arti- edge bases. In this paper, we investigate how to cle mentioning Donald Trump is probably related to politics. keep the freshness of the knowledge base by syn- Moreover, new entities are continuously emerging and most chronizing it with its data source (usually ency- of them are popular, such as iphone 8. However, it is hard for clopedia websites). A direct solution is revisiting a knowledge base to cover these entities in time even if the the whole encyclopedia periodically and rerun the encyclopedia websites have already covered them. entire pipeline of the construction of knowledge freshness base like most existing methods.
    [Show full text]
  • Ontologies and Semantic Web for the Internet of Things - a Survey
    See discussions, stats, and author profiles for this publication at: https://www.researchgate.net/publication/312113565 Ontologies and Semantic Web for the Internet of Things - a survey Conference Paper · October 2016 DOI: 10.1109/IECON.2016.7793744 CITATIONS READS 5 256 2 authors: Ioan Szilagyi Patrice Wira Université de Haute-Alsace Université de Haute-Alsace 10 PUBLICATIONS 17 CITATIONS 122 PUBLICATIONS 679 CITATIONS SEE PROFILE SEE PROFILE Some of the authors of this publication are also working on these related projects: Physics of Solar Cells and Systems View project Artificial intelligence for renewable power generation and management: Application to wind and photovoltaic systems View project All content following this page was uploaded by Patrice Wira on 08 January 2018. The user has requested enhancement of the downloaded file. Ontologies and Semantic Web for the Internet of Things – A Survey Ioan Szilagyi, Patrice Wira MIPS Laboratory, University of Haute-Alsace, Mulhouse, France {ioan.szilagyi; patrice.wira}@uha.fr Abstract—The reality of Internet of Things (IoT), with its one of the most important task in an IoT system [6]. Providing growing number of devices and their diversity is challenging interoperability among the things is “one of the most current approaches and technologies for a smarter integration of fundamental requirements to support object addressing, their data, applications and services. While the Web is seen as a tracking and discovery as well as information representation, convenient platform for integrating things, the Semantic Web can storage, and exchange” [4]. further improve its capacity to understand things’ data and facilitate their interoperability. In this paper we present an There is consensus that Semantic Technologies is the overview of some of the Semantic Web technologies used in IoT appropriate tool to address the diversity of Things [4], [7]–[9].
    [Show full text]
  • Learning Ontologies from RDF Annotations
    /HDUQLQJÃRQWRORJLHVÃIURPÃ5')ÃDQQRWDWLRQV $OH[DQGUHÃ'HOWHLOÃ&DWKHULQHÃ)DURQ=XFNHUÃ5RVHÃ'LHQJ ACACIA project, INRIA, 2004, route des Lucioles, B.P. 93, 06902 Sophia Antipolis, France {Alexandre.Delteil, Catherine.Faron, Rose.Dieng}@sophia.inria.fr $EVWUDFW objects, as in [Mineau, 1990; Carpineto and Romano, 1993; Bournaud HWÃDO., 2000]. In this paper, we present a method for learning Since all RDF annotations are gathered inside a common ontologies from RDF annotations of Web RDF graph, the problem which arises is the extraction of a resources by systematically generating the most description for a given resource from the whole RDF graph. specific generalization of all the possible sets of After a brief description of the RDF data model (Section 2) resources. The preliminary step of our method and of RDF Schema (Section 3), Section 4 presents several consists in extracting (partial) resource criteria for extracting partial resource descriptions. In order descriptions from the whole RDF graph gathering to deal with the intrinsic complexity of the building of a all the annotations. In order to deal with generalization hierarchy, we propose an incremental algorithmic complexity, we incrementally build approach by gradually increasing the size of the descriptions the ontology by gradually increasing the size of the resource descriptions we consider. we consider. The principle of the approach is explained in Section 5 and more deeply detailed in Section 6. Ã ,QWURGXFWLRQ Ã 7KHÃ5')ÃGDWDÃPRGHO The Semantic Web, expected to be the next step that will RDF is the emerging Web standard for annotating resources lead the Web to its full potential, will be based on semantic HWÃDO metadata describing all kinds of Web resources.
    [Show full text]
  • Rdfa in XHTML: Syntax and Processing Rdfa in XHTML: Syntax and Processing
    RDFa in XHTML: Syntax and Processing RDFa in XHTML: Syntax and Processing RDFa in XHTML: Syntax and Processing A collection of attributes and processing rules for extending XHTML to support RDF W3C Recommendation 14 October 2008 This version: http://www.w3.org/TR/2008/REC-rdfa-syntax-20081014 Latest version: http://www.w3.org/TR/rdfa-syntax Previous version: http://www.w3.org/TR/2008/PR-rdfa-syntax-20080904 Diff from previous version: rdfa-syntax-diff.html Editors: Ben Adida, Creative Commons [email protected] Mark Birbeck, webBackplane [email protected] Shane McCarron, Applied Testing and Technology, Inc. [email protected] Steven Pemberton, CWI Please refer to the errata for this document, which may include some normative corrections. This document is also available in these non-normative formats: PostScript version, PDF version, ZIP archive, and Gzip’d TAR archive. The English version of this specification is the only normative version. Non-normative translations may also be available. Copyright © 2007-2008 W3C® (MIT, ERCIM, Keio), All Rights Reserved. W3C liability, trademark and document use rules apply. Abstract The current Web is primarily made up of an enormous number of documents that have been created using HTML. These documents contain significant amounts of structured data, which is largely unavailable to tools and applications. When publishers can express this data more completely, and when tools can read it, a new world of user functionality becomes available, letting users transfer structured data between applications and web sites, and allowing browsing applications to improve the user experience: an event on a web page can be directly imported - 1 - How to Read this Document RDFa in XHTML: Syntax and Processing into a user’s desktop calendar; a license on a document can be detected so that users can be informed of their rights automatically; a photo’s creator, camera setting information, resolution, location and topic can be published as easily as the original photo itself, enabling structured search and sharing.
    [Show full text]
  • The Application of Semantic Web Technologies to Content Analysis in Sociology
    THEAPPLICATIONOFSEMANTICWEBTECHNOLOGIESTO CONTENTANALYSISINSOCIOLOGY MASTER THESIS tabea tietz Matrikelnummer: 749153 Faculty of Economics and Social Science University of Potsdam Erstgutachter: Alexander Knoth, M.A. Zweitgutachter: Prof. Dr. rer. nat. Harald Sack Potsdam, August 2018 Tabea Tietz: The Application of Semantic Web Technologies to Content Analysis in Soci- ology, , © August 2018 ABSTRACT In sociology, texts are understood as social phenomena and provide means to an- alyze social reality. Throughout the years, a broad range of techniques evolved to perform such analysis, qualitative and quantitative approaches as well as com- pletely manual analyses and computer-assisted methods. The development of the World Wide Web and social media as well as technical developments like optical character recognition and automated speech recognition contributed to the enor- mous increase of text available for analysis. This also led sociologists to rely more on computer-assisted approaches for their text analysis and included statistical Natural Language Processing (NLP) techniques. A variety of techniques, tools and use cases developed, which lack an overall uniform way of standardizing these approaches. Furthermore, this problem is coupled with a lack of standards for reporting studies with regards to text analysis in sociology. Semantic Web and Linked Data provide a variety of standards to represent information and knowl- edge. Numerous applications make use of these standards, including possibilities to publish data and to perform Named Entity Linking, a specific branch of NLP. This thesis attempts to discuss the question to which extend the standards and tools provided by the Semantic Web and Linked Data community may support computer-assisted text analysis in sociology. First, these said tools and standards will be briefly introduced and then applied to the use case of constitutional texts of the Netherlands from 1884 to 2016.
    [Show full text]
  • The Semantic Web: the Origins of Artificial Intelligence Redux
    The Semantic Web: The Origins of Artificial Intelligence Redux Harry Halpin ICCS, School of Informatics University of Edinburgh 2 Buccleuch Place Edinburgh EH8 9LW Scotland UK Fax:+44 (0) 131 650 458 E-mail:[email protected] Corresponding author is Harry Halpin. For further information please contact him. This is the tear-off page. To facilitate blind review. Title:The Semantic Web: The Origins of AI Redux working process managed to both halt the fragmentation of Submission for HPLMC-04 the Web and create accepted Web standards through its con- sensus process and its own research team. The W3C set three long-term goals for itself: universal access, Semantic Web, and a web of trust, and since its creation these three goals 1 Introduction have driven a large portion of development of the Web(W3C, 1999) The World Wide Web is considered by many to be the most significant computational phenomenon yet, although even by One comparable program is the Hilbert Program in mathe- the standards of computer science its development has been matics, which set out to prove all of mathematics follows chaotic. While the promise of artificial intelligence to give us from a finite system of axioms and that such an axiom system machines capable of genuine human-level intelligence seems is consistent(Hilbert, 1922). It was through both force of per- nearly as distant as it was during the heyday of the field, the sonality and merit as a mathematician that Hilbert was able ubiquity of the World Wide Web is unquestionable. If any- to set the research program and his challenge led many of the thing it is the Web, not artificial intelligence as traditionally greatest mathematical minds to work.
    [Show full text]
  • Simplicity: Semantics-Sensitive Integrated Matching for Picture Libraries
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 23, NO. 9, SEPTEMBER 2001 1 SIMPLIcity: Semantics-Sensitive Integrated Matching for Picture LIbraries James Z. Wang, Member, IEEE,JiaLi,Member, IEEE, and Gio Wiederhold, Fellow, IEEE AbstractÐThe need for efficient content-based image retrieval has increased tremendously in many application areas such as biomedicine, military, commerce, education, and Web image classification and searching. We present here SIMPLIcity (Semantics- sensitive Integrated Matching for Picture LIbraries), an image retrieval system, which uses semantics classification methods, a wavelet-based approach for feature extraction, and integrated region matching based upon image segmentation. As in other region- based retrieval systems, an image is represented by a set of regions, roughly corresponding to objects, which are characterized by color, texture, shape, and location. The system classifies images into semantic categories, such as textured-nontextured, graph- photograph. Potentially, the categorization enhances retrieval by permitting semantically-adaptive searching methods and narrowing down the searching range in a database. A measure for the overall similarity between images is developed using a region-matching scheme that integrates properties of all the regions in the images. Compared with retrieval based on individual regions, the overall similarity approach 1) reduces the adverse effect of inaccurate segmentation, 2) helps to clarify the semantics of a particular region, and 3) enables a simple querying interface for region-based image retrieval systems. The application of SIMPLIcity to several databases, including a database of about 200,000 general-purpose images, has demonstrated that our system performs significantly better and faster than existing ones. The system is fairly robust to image alterations.
    [Show full text]
  • KBART: Knowledge Bases and Related Tools
    NISO-RP-9-2010 KBART: Knowledge Bases and Related Tools A Recommended Practice of the National Information Standards Organization (NISO) and UKSG Prepared by the NISO/UKSG KBART Working Group January 2010 i About NISO Recommended Practices A NISO Recommended Practice is a recommended "best practice" or "guideline" for methods, materials, or practices in order to give guidance to the user. Such documents usually represent a leading edge, exceptional model, or proven industry practice. All elements of Recommended Practices are discretionary and may be used as stated or modified by the user to meet specific needs. This recommended practice may be revised or withdrawn at any time. For current information on the status of this publication contact the NISO office or visit the NISO website (www.niso.org). Published by National Information Standards Organization (NISO) One North Charles Street, Suite 1905 Baltimore, MD 21201 www.niso.org Copyright © 2010 by the National Information Standards Organization and the UKSG. All rights reserved under International and Pan-American Copyright Conventions. For noncommercial purposes only, this publication may be reproduced or transmitted in any form or by any means without prior permission in writing from the publisher, provided it is reproduced accurately, the source of the material is identified, and the NISO/UKSG copyright status is acknowledged. All inquires regarding translations into other languages or commercial reproduction or distribution should be addressed to: NISO, One North Charles Street,
    [Show full text]
  • Exploiting Semantic Web Knowledge Graphs in Data Mining
    Exploiting Semantic Web Knowledge Graphs in Data Mining Inauguraldissertation zur Erlangung des akademischen Grades eines Doktors der Naturwissenschaften der Universit¨atMannheim presented by Petar Ristoski Mannheim, 2017 ii Dekan: Dr. Bernd Lübcke, Universität Mannheim Referent: Professor Dr. Heiko Paulheim, Universität Mannheim Korreferent: Professor Dr. Simone Paolo Ponzetto, Universität Mannheim Tag der mündlichen Prüfung: 15 Januar 2018 Abstract Data Mining and Knowledge Discovery in Databases (KDD) is a research field concerned with deriving higher-level insights from data. The tasks performed in that field are knowledge intensive and can often benefit from using additional knowledge from various sources. Therefore, many approaches have been proposed in this area that combine Semantic Web data with the data mining and knowledge discovery process. Semantic Web knowledge graphs are a backbone of many in- formation systems that require access to structured knowledge. Such knowledge graphs contain factual knowledge about real word entities and the relations be- tween them, which can be utilized in various natural language processing, infor- mation retrieval, and any data mining applications. Following the principles of the Semantic Web, Semantic Web knowledge graphs are publicly available as Linked Open Data. Linked Open Data is an open, interlinked collection of datasets in machine-interpretable form, covering most of the real world domains. In this thesis, we investigate the hypothesis if Semantic Web knowledge graphs can be exploited as background knowledge in different steps of the knowledge discovery process, and different data mining tasks. More precisely, we aim to show that Semantic Web knowledge graphs can be utilized for generating valuable data mining features that can be used in various data mining tasks.
    [Show full text]
  • The Ontology Web Language (OWL) for a Multi-Agent Understating
    The Ontology Web Language (OWL) for a Multi- Agent Understating System Mostafa M. Aref Zhengbo Zhou Department of Computer Science and Engineering University of Bridgeport, Bridgeport, CT 06601 email: [email protected] Abstract— Computer understanding is a challenge OWL is a Web Ontology Language. It is built on top of problem in Artificial Intelligence. A multi-agent system has RDF – Resource Definition Framework and written in XML. been developed to tackle this problem. Among its modules is It is a part of Semantic Web Vision, and is designed to be its knowledge base (vocabulary agents). This paper interpreted by computers, not for being read by people. discusses the use of the Ontology Web Language (OWL) to OWL became a W3C (World Wide Web Consortium) represent the knowledge base. An example of applying OWL Recommendation in February 2004 [2]. The OWL is a in sentence understanding is given. Followed by an language for defining and instantiating Web ontologies. evaluation of OWL. OWL ontology may include the descriptions of classes, properties, and their instances [3]. Given such ontology, the OWL formal semantics specifies how to derive its logical 1. INTRODUCTION consequences, i.e. facts not literally present in the ontology, but entailed by the semantics. One of various definitions for Artificial Intelligence is “The study of how to make computers do things which, at the Section 2 describes a multi-agents understanding system. moment, people do better”[7]. From the definition of AI Section 3 gives a brief description of a newly standardized mentioned above, “Understanding” can be looked as the technique, Web Ontology Language—OWL.
    [Show full text]
  • Semantic Web
    Semantic Web Ing. Federico Chesani Corso di Fondamenti di Intelligenza Artificiale M Outline 1. Introduction a) The map of the Web (accordingly to Tim Berners-Lee) b) The current Web and its limits c) The Semantic Web idea 2. Semantic Information (a bird’s eye view) a) Semantic Models b) Ontologies c) Few examples 3. Semantic Web Tools a) Unique identifiers -URI b) XML c) RDF and SPARQL d) OWL 4. Semantic Web: where are we? a) Problems against the success of SW proposal b) Critics against SW c) Few considerations d) Few links to start with The Web Map (by Berners-Lee) ©Tim Berners-Lee, http://www.w3.org/2007/09/map/main.jpg About the content Knowledge Representation Semantic Web Web The Web 1.0 … • Information represented by means of: – Natural language – Images, multimedia, graphic rendering/aspect • Human Users easily exploit all this means for: – Deducting facts from partial information – Creating mental asociations (between the facts and, e.g., the images) – They use different communication channels at the same time (contemporary use of many primitive senses) The Web 1.0 … • The content is published on the web with the principal aim of being “human-readable” – Standard HTML is focused on how to represent the content – There is no notion of what is represented – Few tags (e.g. <title>) provide an implicit semantics but … • … their content is not structured • … their use is not really standardized The Web 1.0 … We can identify the title by means of its representation (<h1>, <b>) … … what if tomorrow the designer changes the format of the web pages? <h1> <!-- inizio TITOLO --> <B> Finanziaria, il voto slitta a domani<br> Al Senato va in scena l&#039;assurdo </B> <!-- fine TITOLO --> </h1> The Web 1.0 … • Web pages contain also links to other pages, but ..
    [Show full text]