The Relation Between Ontologies and XML Schemata
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
OWL 2 Web Ontology Language Quick Reference Guide
OWL 2 Web Ontology Language Quick Reference W3C Proposed Recommendation 22 September Guide 2009 OWL 2 Web Ontology Language Quick Reference Guide W3C Proposed Recommendation 22 September 2009 This version: http://www.w3.org/TR/2009/PR-owl2-quick-reference-20090922/ Latest version: http://www.w3.org/TR/owl2-quick-reference/ Previous version: http://www.w3.org/TR/2009/WD-owl2-quick-reference-20090611/ (color-coded diff) Editors: Jie Bao, Rensselaer Polytechnic Institute Elisa F. Kendall, Sandpiper Software, Inc. Deborah L. McGuinness, Rensselaer Polytechnic Institute Peter F. Patel-Schneider, Bell Labs Research, Alcatel-Lucent Contributors: Li Ding, Rensselaer Polytechnic Institute Ankesh Khandelwal, Rensselaer Polytechnic Institute This document is also available in these non-normative formats: PDF version, Reference Card. Copyright © 2009 W3C® (MIT, ERCIM, Keio), All Rights Reserved. W3C liability, trademark and document use rules apply. Abstract The OWL 2 Web Ontology Language, informally OWL 2, is an ontology language for the Semantic Web with formally defined meaning. OWL 2 ontologies provide classes, properties, individuals, and data values and are stored as Semantic Web documents. OWL 2 ontologies can be used along with information written in RDF, and OWL 2 ontologies themselves are primarily exchanged as RDF documents. The OWL 2 Document Overview describes the overall state of OWL 2, and should be read before other OWL 2 documents. Page 1 of 15 http://www.w3.org/TR/2009/PR-owl2-quick-reference-20090922/ OWL 2 Web Ontology Language Quick Reference W3C Proposed Recommendation 22 September Guide 2009 This document provides a non-normative quick reference guide to the OWL 2 language. -
Bi-Directional Transformation Between Normalized Systems Elements and Domain Ontologies in OWL
Bi-directional Transformation between Normalized Systems Elements and Domain Ontologies in OWL Marek Suchanek´ 1 a, Herwig Mannaert2, Peter Uhnak´ 3 b and Robert Pergl1 c 1Faculty of Information Technology, Czech Technical University in Prague, Thakurova´ 9, Prague, Czech Republic 2Normalized Systems Institute, University of Antwerp, Prinsstraat 13, Antwerp, Belgium 3NSX bvba, Wetenschapspark Universiteit Antwerpen, Galileilaan 15, 2845 Niel, Belgium Keywords: Ontology, Normalized Systems, Transformation, Model-driven Development, Ontology Engineering, Software Modelling. Abstract: Knowledge representation in OWL ontologies gained a lot of popularity with the development of Big Data, Artificial Intelligence, Semantic Web, and Linked Open Data. OWL ontologies are very versatile, and there are many tools for analysis, design, documentation, and mapping. They can capture concepts and categories, their properties and relations. Normalized Systems (NS) provide a way of code generation from a model of so-called NS Elements resulting in an information system with proven evolvability. The model used in NS contains domain-specific knowledge that can be represented in an OWL ontology. This work clarifies the potential advantages of having OWL representation of the NS model, discusses the design of a bi-directional transformation between NS models and domain ontologies in OWL, and describes its implementation. It shows how the resulting ontology enables further work on the analytical level and leverages the system design. Moreover, due to the fact that NS metamodel is metacircular, the transformation can generate ontology of NS metamodel itself. It is expected that the results of this work will help with the design of larger real-world applications as well as the metamodel and that the transformation tool will be further extended with additional features which we proposed. -
SLA Information Technology Division Metadata for Video: Too Much
3/1/2019 Metadata for Video: Too Much Content, Not Enough Information | SLA Information Technology Division Home About Us » Events » Sections » Enter search keyword SLA Information Technology Division Awards » Making Edgier Easier. We're IT! Current b/ITe (v31n5) » b/ITe Archives Virtual Events What’s New Categorized | Uncategorized Metadata for Video: Too Much Content, Not Enough Information Posted on 31 August 2012. by Wayne Pender, McGill University, 2012 Joe Ann Clifton Student Award Winner (Published September 1, 2012) Abstract Television news libraries have struggled with cataloguing, organizing and storing visual materials for efficient search and retrieval for years. This task has been complicated by the emergence of digital video and the exponential growth in the holdings of television news libraries. In parallel, the ability for non-professionals to shoot, edit and publish videos of increasing production value and complexity on the web has flooded the Internet with videos that appeal to a wide audience and subject interest. The present survey looks at the metadata protocols and practices in place for an internal audience in professional operations and on display for the general public on the Internet. The study finds that the lack of a common metadata schema can make much of the material inaccessible. Literature pertaining to this area is reviewed and future direction is discussed. http://it.sla1.org/2012/08/metadata/ 1/10 3/1/2019 Metadata for Video: Too Much Content, Not Enough Information | SLA Information Technology Division Keywords: metadata, video, XML, RDF, MXF, television, news, YouTube Paper Searching and retrieving visual content is problematic. The search for specific moving images on film and video has long been a task for television news professionals, but now with wide spread availability of video resources on the Internet it has become a task for anyone. -
Instruction for Using XML Notepad
Using XML Notepad to Read, Edit, and Parse FGDC-CSDGM XML Metadata. Currently many tools exist that allow a user to work with XML metadata files. However, most of the main metadata creation/editing tools do not provide a means of validating a metadata record for compliance with the FGDC-CSDGM standard or its variants, such as the Biological Data Profile (BDP) standard. One of the most reliable ways to validate metadata files (that is, check files for completion and/or errors) is with the USGS Metadata Parser (MP) utilities. The MP tool and the other tools distributed in the package were developed by Peter Schweitzer and are freely available online (http://geology.usgs.gov/tools/metadata/tools/doc/mp.html). Once configured properly, MP can process metadata files and be used to produce a text file with a list of all errors found. A user can then use this list of errors to find and correct the problems using a text editor or the metadata editor of their choice. The challenge with this method is that finding and correcting the error within a metadata editor based solely on the error message from MP can be confusing and frustrating. Additionally, if there are many errors or complex problems to correct, the process will require multiple iterations of running MP and subsequently correcting errors in a metadata editor. All in all, the process can be fairly time consuming and onerous, especially for those without a detailed understanding of the FGDC-CSDGM standard. The methodology for metadata editing and validation described in this document tries to address some of these problems with an alternative approach. -
Cisco XML Schemas
CHAPTER 14 Cisco XML Schemas This chapter contains information about common XML schemas. The structure and allowable content of the XML request and response instances supported by the Cisco IOS XR XML application programming interface (API) are documented by means of XML schemas (.xsd files). The XML schemas are documented using the standard World Wide Web Consortium (W3C) XML schema language, which provides a much more powerful and flexible mechanism for describing schemas than can be achieved using Document Type Definitions (DTDs). The set of XML schemas consists of a small set of common high-level schemas and a larger number of component-specific schemas as described in this chapter. For more information on the W3C XML Schema standard, see this URL: http://www.w3.org/XML/Schema This chapter contains these sections: • XML Schema Retrieval, page 14-135 • Common XML Schemas, page 14-136 • Component XML Schemas, page 14-136 XML Schema Retrieval The XML schemas that belong to the features in a particular package are obtained as a .tar file from cisco.com. To retrieve the XML schemas, you must: 1. Click this URL to display the Downloads page: http://tools.cisco.com/support/downloads/go/Redirect.x?mdfid=268437899 Note Select Downloads. Only customer or partner viewers can access the Download Software page. Guest users will get an error. 2. Select Cisco IOS XR Software. 3. Select IOS XR XML Schemas. 4. Select the XML schema for your platform. Once untarred, all the XML schema files appear as a flat directory of .xsd files and can be opened with any XML schema viewing application, such as XMLSpy. -
Modularization of XHTML in XML Schema Modularization of XHTML™ in XML Schema
Modularization of XHTML in XML Schema Modularization of XHTML™ in XML Schema Modularization of XHTML™ in XML Schema W3C Working Draft - 22 March 2001 This version: http://www.w3.org/TR/2001/WD-xhtml-m12n-schema-20010322 (Single HTML file [p.1] , PostScript version, PDF version, ZIP archive, or Gzip’d TAR archive) Latest version: http://www.w3.org/TR/xhtml-m12n-schema Editors: Daniel Austin, Mozquito Technologies AG Shane McCarron, Applied Testing and Technology Copyright ©2001 W3C® (MIT, INRIA, Keio), All Rights Reserved. W3C liability, trademark, document use and software licensing rules apply. Abstract This document describes a methodology for the modularization of XHTML using XML Schema. Modularization of XHTML allows document authors to modify and extend XHTML in a conformant way. Status of This Document This section describes the status of this document at the time of its publication. Other documents may supersede this document. The latest status of this document series is maintained at the W3C. This is the first public "Working Draft" of "Modularization of XHTML in XML Schema" for review by members of the W3C and other interested parties in the general public. It is a stand-alone document to ease its review. Once the methodology described in this document become mature, it will be integrated into a future document forthcoming from the HTML Working Group. This document is still in its early stage, and may be updated, replaced, or obsoleted by other documents at any time. Publication of this Working Draft does not imply endorsement by the W3C, and it is inappropriate to use W3C Working Drafts as reference material or to cite them as other than "work in progress". -
The Development of Algorithms for On-Demand Map Editing for Internet and Mobile Users with Gml and Svg
THE DEVELOPMENT OF ALGORITHMS FOR ON-DEMAND MAP EDITING FOR INTERNET AND MOBILE USERS WITH GML AND SVG Miss. Ida K.L CHEUNG a, , Mr. Geoffrey Y.K. SHEA b a Department of Land Surveying & Geo-Informatics, The Hong Kong Polytechnic University, Hung Hom, Kowloon, Hong Kong, email: [email protected] b Department of Land Surveying & Geo-Informatics, The Hong Kong Polytechnic University, Hung Hom, Kowloon, Hong Kong, email: [email protected] Commission VI, PS WG IV/2 KEY WORDS: Spatial Information Sciences, GIS, Research, Internet, Interoperability ABSTRACT: The widespread availability of the World Wide Web has led to a rapid increase in the amount of data accessing, sharing and disseminating that gives the opportunities for delivering maps over the Internet as well as small mobile devices. In GIS industry, many vendors or companies have produced their own web map products which have their own version, data model and proprietary data formats without standardization. Such problem has long been an issue. Therefore, Geographic Markup Language (GML) was designed to provide solutions. GML is an XML grammar written in XML Schema for the modelling, transport, and storage of geographic information including both spatial and non-spatial properties of geographic features. GML is developed by Open GIS Consortium in order to promote spatial data interoperability and open standard. Since GML is still a newly developed standard, this promising research field provides a standardized method to integrate web-based mapping information in terms of data modelling, spatial data representation mechanism and graphic presentation. As GML is not for data display, SVG is an ideal vector graphic for displaying geographic data. -
Data Models for Home Services
__________________________________________PROCEEDING OF THE 13TH CONFERENCE OF FRUCT ASSOCIATION Data Models for Home Services Vadym Kramar, Markku Korhonen, Yury Sergeev Oulu University of Applied Sciences, School of Engineering Raahe, Finland {vadym.kramar, markku.korhonen, yury.sergeev}@oamk.fi Abstract An ultimate penetration of communication technologies allowing web access has enriched a conception of smart homes with new paradigms of home services. Modern home services range far beyond such notions as Home Automation or Use of Internet. The services expose their ubiquitous nature by being integrated into smart environments, and provisioned through a variety of end-user devices. Computational intelligence require a use of knowledge technologies, and within a given domain, such requirement as a compliance with modern web architecture is essential. This is where Semantic Web technologies excel. A given work presents an overview of important terms, vocabularies, and data models that may be utilised in data and knowledge engineering with respect to home services. Index Terms: Context, Data engineering, Data models, Knowledge engineering, Semantic Web, Smart homes, Ubiquitous computing. I. INTRODUCTION In recent years, a use of Semantic Web technologies to build a giant information space has shown certain benefits. Rapid development of Web 3.0 and a use of its principle in web applications is the best evidence of such benefits. A traditional database design in still and will be widely used in web applications. One of the most important reason for that is a vast number of databases developed over years and used in a variety of applications varying from simple web services to enterprise portals. In accordance to Forrester Research though a growing number of document, or knowledge bases, such as NoSQL is not a hype anymore [1]. -
Integrating Fuzzy Logic in Ontologies
INTEGRATING FUZZY LOGIC IN ONTOLOGIES Silvia Calegari and Davide Ciucci Dipartimento di Informatica, Sistemistica e Comunicazione, Universita` degli Studi di Milano Bicocca, via Bicocca degli Arcimboldi 8, 20126 Milano, Italy Keywords: Concept modifiers, fuzzy logics, fuzzy ontologies, membership modifiers, KAON, ontology editor. Abstract: Ontologies have proved to be very useful in sharing concepts across applications in an unambiguous way. Nowadays, in ontology-based applications information is often vague and imprecise. This is a well-known problem especially for semantics-based applications, such as e-commerce, knowledge management, web por- tals, etc. In computer-aided reasoning, the predominant paradigm to manage vague knowledge is fuzzy set theory. This paper presents an enrichment of classical computational ontologies with fuzzy logic to create fuzzy ontologies. So, it is a step towards facing the nuances of natural languages with ontologies. Our pro- posal is developed in the KAON ontology editor, that allows to handle ontology concepts in an high-level environment. 1 INTRODUCTION A possible solution to handle uncertain data and, hence, to tackle these problems, is to incorporate fuzzy logic into ontologies. The aim of fuzzy set An ontology is a formal conceptualization of a partic- theory (Klir and Yuan, 1995) introduced by L. A. ular domain of interest shared among heterogeneous Zadeh (Zadeh, 1965) is to describe vague concepts applications. It consists of entities, attributes, rela- through a generalized notion of set, according to tionships and axioms to provide a common under- which an object may belong to a certain degree (typ- standing of the real world (Lammari and Mtais, 2004; ically a real number from the interval [0,1]) to a set. -
Determining the Output Schema of an XSLT Stylesheet
Determining the Output Schema of an XSLT Stylesheet Sven Groppe and Jinghua Groppe University of Innsbruck, Technikerstrasse 21a, A-6020 Innsbruck, Austria {Sven.Groppe, Jinghua Groppe}@uibk.ac.at Abstract. The XSLT language is used to describe transformations of XML documents into other formats. The transformed XML documents conform to output schemas of the used XSLT stylesheet. Output schemas of XSLT stylesheets can be used for a static analysis of the used XSLT stylesheet, to automatically detect the XSLT stylesheet, which has been used for the trans- formation, of target XML documents or to reason on the output schema without access to the target XML documents. In this paper, we describe how to auto- matically determine such an output schema of a given XSLT stylesheet, where we only consider XML to XML transformations. The input of our proposed output schema generator is the XSLT stylesheet and the schema of the input XML documents. The experimental evaluation shows that our prototype can de- termine the output schemas of nearly all typical XSLT stylesheets. 1 Introduction Among other usages of XML, XML is the most widely used data model for exchang- ing data on the web and elsewhere. For the exchange of data, we have to transform the data from one format into another format whenever the two exchange partners use different formats. The exchange partners can use different formats, which might be a proprietary company standard, a proprietary application format or other standard for- mats, for historical, political or other reasons. We focus on XSLT [23] as transforma- tion language for the XML data. -
XHTML+Rdfa 1.1 - Third Edition Table of Contents
XHTML+RDFa 1.1 - Third Edition Table of Contents XHTML+RDFa 1.1 - Third Edition Support for RDFa via XHTML Modularization W3C Recommendation 17 March 2015 This version: http://www.w3.org/TR/2015/REC-xhtml-rdfa-20150317/ Latest published version: http://www.w3.org/TR/xhtml-rdfa/ Implementation report: http://www.w3.org/2010/02/rdfa/wiki/CR-ImplementationReport Previous version: http://www.w3.org/TR/2014/PER-xhtml-rdfa-20141216/ Previous Recommendation: http://www.w3.org/TR/2013/REC-xhtml-rdfa-20130822/ Editor: Shane McCarron, Applied Testing and Technology, Inc., [email protected] Please check the errata for any errors or issues reported since publication. This document is also available in these non-normative formats: XHTML+RDFa, Diff from Previous Recommendation, Postscript version, and PDF version The English version of this specification is the only normative version. Non-normative translations may also be available. Copyright © 2007-2015 W3C® (MIT, ERCIM, Keio, Beihang). W3C liability, trademark and document use rules apply. Abstract RDFa Core 1.1 [RDFA-CORE [p.61] ] defines attributes and syntax for embedding semantic markup in Host Languages. This document defines one such Host Language. This language is a superset of XHTML 1.1 [XHTML11-2e [p.61] ], integrating the attributes as defined in RDFa Core 1.1. This document is intended for authors who want to create XHTML Family documents that embed rich semantic markup. - 1 - Status of This Document XHTML+RDFa 1.1 - Third Edition Status of This Document This section describes the status of this document at the time of its publication. -
Arxiv:1009.3391V3 [Cs.LO] 4 Nov 2010
Fuzzy Ontology Representation using OWL 2 I Fernando Bobilloa, Umberto Stracciab aDepartment of Computer Science and Systems Engineering, University of Zaragoza, Spain bIstituto di Scienza e Tecnologie dell'Informazione (ISTI - CNR), Pisa, Italy Abstract The need to deal with vague information in Semantic Web languages is rising in importance and, thus, calls for a standard way to represent such information. We may address this issue by either extending current Semantic Web languages to cope with vagueness, or by providing a procedure to represent such informa- tion within current standard languages and tools. In this work, we follow the latter approach, by identifying the syntactic differences that a fuzzy ontology language has to cope with, and by proposing a concrete methodology to rep- resent fuzzy ontologies using OWL 2 annotation properties. We also report on the prototypical implementations. Key words: Fuzzy OWL 2, Fuzzy Ontologies, Fuzzy Languages for the Semantic Web, Fuzzy Description Logics 1. Introduction Today, there is a growing interest in the development of knowledge representa- tion formalisms able to deal with uncertainty, which is a very common require- ment in real world applications. Despite the undisputed success of ontologies, classical ontology languages are not appropriate to deal with vagueness or impre- cision in the knowledge, which is inherent to most of the real world application domains [29]. Since fuzzy set theory and fuzzy logic [30] are suitable formalisms to handle these types of knowledge, fuzzy ontologies emerge as useful in several applica- arXiv:1009.3391v3 [cs.LO] 4 Nov 2010 tions, ranging from (multimedia) information retrieval to image interpretation, ontology mapping, matchmaking, decision making, or the Semantic Web [19].