Redalyc.Cognitive Modules of an NLP Knowledge Base for Language

Total Page:16

File Type:pdf, Size:1020Kb

Redalyc.Cognitive Modules of an NLP Knowledge Base for Language Procesamiento del Lenguaje Natural ISSN: 1135-5948 [email protected] Sociedad Española para el Procesamiento del Lenguaje Natural España Periñán-Pascual, Carlos; Arcas-Túnez, Francisco Cognitive Modules of an NLP Knowledge Base for Language Understanding Procesamiento del Lenguaje Natural, núm. 39, 2007, pp. 197-204 Sociedad Española para el Procesamiento del Lenguaje Natural Jaén, España Available in: http://www.redalyc.org/articulo.oa?id=515751739024 How to cite Complete issue Scientific Information System More information about this article Network of Scientific Journals from Latin America, the Caribbean, Spain and Portugal Journal's homepage in redalyc.org Non-profit academic project, developed under the open access initiative Procesamiento del Lenguaje Natural, nº39 (2007), pp. 197-204 recibido 30-04-2007; aceptado 22-06-2007 Cognitive Modules of an NLP Knowledge Base for Language Understanding Carlos Periñán-Pascual Francisco Arcas-Túnez Universidad Católica San Antonio Universidad Católica San Antonio Campus de los Jerónimos s/n Campus de los Jerónimos s/n 30107 Guadalupe - Murcia (Spain) 30107 Guadalupe - Murcia (Spain) [email protected] [email protected] Resumen : Algunas aplicaciones del procesamiento del lenguaje natural, p.ej. la traducción automática, requieren una base de conocimiento provista de representaciones conceptuales que puedan reflejar la estructura del sistema cognitivo del ser humano. En cambio, tareas como la indización automática o la extracción de información pueden ser realizadas con una semántica superficial. De todos modos, la construcción de una base de conocimiento robusta garantiza su reutilización en la mayoría de las tareas del procesamiento del lenguaje natural. El propósito de este artículo es describir los principales módulos cognitivos de FunGramKB, una base de conocimiento léxico-conceptual multipropósito para su implementación en sistemas del procesamiento del lenguaje natural. Palabras clave : Representación del conocimiento, ontología, razonamiento, postulado de significado. Abstract : Some natural language processing systems, e.g. machine translation, require a knowledge base with conceptual representations reflecting the structure of human beings’ cognitive system. In some other systems, e.g. automatic indexing or information extraction, surface semantics could be sufficient, but the construction of a robust knowledge base guarantees its use in most natural language processing tasks, consolidating thus the concept of resource reuse. The objective of this paper is to describe FunGramKB, a multipurpose lexico- conceptual knowledge base for natural language processing systems. Particular attention will be paid to the two main cognitive modules, i.e. the ontology and the cognicon. Keywords : Knowledge representation, ontology, reasoning, meaning postulate. functional model in some important aspects 1 FunGramKB with the aim of building a more robust knowledge base. FunGramKB Suite 1 is a user-friendly On the one hand, FunGramKB is environment for the semiautomatic construction multipurpose in the sense that it is both of a multipurpose lexico-conceptual knowledge multifunctional and multilanguage. In other base for a natural language processing (NLP) words, FunGramKB can be reused in various system within the theoretical model of S.C. NLP tasks (e.g. information retrieval and Dik’s Functional Grammar (1978, 1989, 1997). extraction, machine translation, dialogue-based FunGramKB is not a literal implementation of systems, etc) and with several natural Dik’s lexical database, but we depart from the languages. 2 1 We use the name ‘FunGramKB Suite’ to refer to our knowledge engineering tool and 2 English, Spanish, German, French and Italian ‘FunGramKB’ to the resulting knowledge base. are supported in the current version of FunGramKB. ISSN: 1135-5948 © 2007 Sociedad Española para el Procesamiento del Lenguaje Natural Carlos Periñan-Pascual y Francisco Arcas-Túnez On the other hand, our knowledge base is different types (Tulving, 1985): lexico-conceptual, because it comprises two general levels of information: a lexical level • Semantic knowledge, which stores and a cognitive level. In turn, these levels are cognitive information about words; it is made up of several independent but interrelated a kind of mental dictionary. components: • Procedural knowledge, which stores information about how events are Lexical level (i.e. linguistic knowledge): performed in ordinary situations—e.g. • The lexicon stores morphosyntactic, how to ride a bicycle, how to fry an pragmatic and collocational egg...; it is a kind of manual for information of words. everyday actions. • The morphicon helps our system to • Episodic knowledge, which stores handle cases of inflectional information about specific biographic morphology. events or situations—e.g. our wedding- day; it is a kind of personal scrapbook. Cognitive level (i.e. non-linguistic knowledge): Therefore, if there are three types of • The ontology is presented as a knowledge involved in human reasoning, there hierarchical structure of all the concepts must be three different kinds of knowledge that a person has in mind when talking schemata. These schemata are successfully about everyday situations. mapped in an integrated way into the cognitive • The cognicon stores procedural component of FunGramKB: knowledge by means of cognitive macrostructures, i.e. script-like • Semantic knowledge is represented in schemata in which a sequence of the form of meaning postulates in the stereotypical actions is organised on the ontology. basis of temporal continuity, and more • Procedural knowledge is represented in particularly on James Allen's temporal the form of cognitive macrostructures model (Allen, 1983, 1991; Allen and in the cognicon. Ferguson, 1994). • Episodic knowledge can be stored as a • The onomasticon stores information case base. 3 about instances of entities, such as people, cities, products, etc. A key factor for successful reasoning is that all these knowledge schemata (i.e. meaning The main consequence of this two-level postulates, cognitive macrostructures and cases) design is that every lexical module is language- must be represented through the same formal dependent, while every cognitive module is language, so that information sharing could take shared by all languages. In other words, place effectively among all cognitive modules. computational lexicographers must develop one Our formal language is partially founded on lexicon and one morphicon for English, one Dik’s model of semantic representation (1978, lexicon and one morphicon for Spanish and so 1989, 1997), which was initially devised for on, but knowledge engineers build just one machine translation (Connolly and Dik, 1989). ontology, one cognicon and one onomasticon to Computationally speaking, when storing process any language input cognitively. Section cognitive knowledge through FunGramKB 2 gives a brief account on the psychological Suite, a syntactic-semantic checker is triggered, foundation of FunGramKB cognitive level, and so that consistent well-formed constructs can be sections 3 and 4 describe the two main stored. Moreover, a parser outputs an XML- cognitive modules in that level, i.e. the formatted feature-value structure used as the ontology and the cognicon. input for the reasoning engine, so that 2 Cognitive knowledge in natural language understanding 3 FunGramKB can be very useful in case-based In cognitive psychology, common-sense reasoning, where problems are solved by knowledge is usually divided into three remembering previous similar cases and reusing general knowledge. 198 Cognitive Modules of an NLP Knowledge Base for Language Understanding inheritance and inference mechanisms can be establishes a high degree of connectivity among applied. Both the syntactic-semantic validator conceptual units by taking into account of meaning postulates and the XML parser were semantic components which are shared by their written in C#. meaning postulates. In order to incorporate human beings’ commonsense, our ontology 3 FunGramKB ontology must identify the relations which can be established among conceptual units, and hence Nowadays there is no single right methodology among lexical units. However, displaying for ontology development. Ontology design semantic similarities and differences through tends to be a creative process, so it is probable taxonomic relations turns out to be more that two ontologies designed by different chaotic than through meaning postulates linked people have a different structuring (Noy and to conceptual units. McGuinness, 2001). To avoid this problem, the ontology model should be founded on a solid 3.3 Three-layered ontological model methodology. The remaining of this section describes five methodological criteria applied to FunGramKB ontology distinguishes three FunGramKB ontology, some of which are different conceptual levels, each one of them based on principles implemented in other NLP with concepts of a different type: metaconcepts, projects (Bouaud et al., 1995; Mahesh, 1996; basic concepts and terminals. Figure (1) Noy and McGuinness, 2001). The definition of illustrates these three types of concepts. these criteria in the analysis and design phases of the ontology model and the strict application #ENTITY of these guidelines in the development phase → #PHYSICAL contributed to avoid some common errors in conceptual modelling. → #OBJECT → #SELF_CONNECTED_OBJECT 3.1 Symbiosis between universality and → +ARTIFICIAL_OBJECT linguistic
Recommended publications
  • A Survey of Top-Level Ontologies to Inform the Ontological Choices for a Foundation Data Model
    A survey of Top-Level Ontologies To inform the ontological choices for a Foundation Data Model Version 1 Contents 1 Introduction and Purpose 3 F.13 FrameNet 92 2 Approach and contents 4 F.14 GFO – General Formal Ontology 94 2.1 Collect candidate top-level ontologies 4 F.15 gist 95 2.2 Develop assessment framework 4 F.16 HQDM – High Quality Data Models 97 2.3 Assessment of candidate top-level ontologies F.17 IDEAS – International Defence Enterprise against the framework 5 Architecture Specification 99 2.4 Terminological note 5 F.18 IEC 62541 100 3 Assessment framework – development basis 6 F.19 IEC 63088 100 3.1 General ontological requirements 6 F.20 ISO 12006-3 101 3.2 Overarching ontological architecture F.21 ISO 15926-2 102 framework 8 F.22 KKO: KBpedia Knowledge Ontology 103 4 Ontological commitment overview 11 F.23 KR Ontology – Knowledge Representation 4.1 General choices 11 Ontology 105 4.2 Formal structure – horizontal and vertical 14 F.24 MarineTLO: A Top-Level 4.3 Universal commitments 33 Ontology for the Marine Domain 106 5 Assessment Framework Results 37 F. 25 MIMOSA CCOM – (Common Conceptual 5.1 General choices 37 Object Model) 108 5.2 Formal structure: vertical aspects 38 F.26 OWL – Web Ontology Language 110 5.3 Formal structure: horizontal aspects 42 F.27 ProtOn – PROTo ONtology 111 5.4 Universal commitments 44 F.28 Schema.org 112 6 Summary 46 F.29 SENSUS 113 Appendix A F.30 SKOS 113 Pathway requirements for a Foundation Data F.31 SUMO 115 Model 48 F.32 TMRM/TMDM – Topic Map Reference/Data Appendix B Models 116 ISO IEC 21838-1:2019
    [Show full text]
  • Chaudron: Extending Dbpedia with Measurement Julien Subercaze
    Chaudron: Extending DBpedia with measurement Julien Subercaze To cite this version: Julien Subercaze. Chaudron: Extending DBpedia with measurement. 14th European Semantic Web Conference, Eva Blomqvist, Diana Maynard, Aldo Gangemi, May 2017, Portoroz, Slovenia. hal- 01477214 HAL Id: hal-01477214 https://hal.archives-ouvertes.fr/hal-01477214 Submitted on 27 Feb 2017 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Chaudron: Extending DBpedia with measurement Julien Subercaze1 Univ Lyon, UJM-Saint-Etienne, CNRS Laboratoire Hubert Curien UMR 5516, F-42023, SAINT-ETIENNE, France [email protected] Abstract. Wikipedia is the largest collaborative encyclopedia and is used as the source for DBpedia, a central dataset of the LOD cloud. Wikipedia contains numerous numerical measures on the entities it describes, as per the general character of the data it encompasses. The DBpedia In- formation Extraction Framework transforms semi-structured data from Wikipedia into structured RDF. However this extraction framework of- fers a limited support to handle measurement in Wikipedia. In this paper, we describe the automated process that enables the creation of the Chaudron dataset. We propose an alternative extraction to the tra- ditional mapping creation from Wikipedia dump, by also using the ren- dered HTML to avoid the template transclusion issue.
    [Show full text]
  • Open Mind Common Sense: Knowledge Acquisition from the General Public
    Open Mind Common Sense: Knowledge Acquisition from the General Public Push Singh, Grace Lim, Thomas Lin, Erik T. Mueller Travell Perkins, Mark Tompkins, Wan Li Zhu MIT Media Laboratory 20 Ames Street Cambridge, MA 02139 USA {push, glim, tlin, markt, wlz}@mit.edu, [email protected], [email protected] Abstract underpinnings for commonsense reasoning (Shanahan Open Mind Common Sense is a knowledge acquisition 1997), there has been far less work on finding ways to system designed to acquire commonsense knowledge from accumulate the knowledge to do so in practice. The most the general public over the web. We describe and evaluate well-known attempt has been the Cyc project (Lenat 1995) our first fielded system, which enabled the construction of which contains 1.5 million assertions built over 15 years a 400,000 assertion commonsense knowledge base. We at the cost of several tens of millions of dollars. then discuss how our second-generation system addresses Knowledge bases this large require a tremendous effort to weaknesses discovered in the first. The new system engineer. With the exception of Cyc, this problem of scale acquires facts, descriptions, and stories by allowing has made efforts to study and build commonsense participants to construct and fill in natural language knowledge bases nearly non-existent within the artificial templates. It employs word-sense disambiguation and intelligence community. methods of clarifying entered knowledge, analogical inference to provide feedback, and allows participants to validate knowledge and in turn each other. Turning to the general public 1 In this paper we explore a possible solution to this Introduction problem of scale, based on one critical observation: Every We would like to build software agents that can engage in ordinary person has common sense of the kind we want to commonsense reasoning about ordinary human affairs.
    [Show full text]
  • Automatic Affective Feedback in an Email Browser
    Automatic Affective Feedback in an Email Browser Hugo Liu Henry Lieberman Ted Selker Software Agents Group Software Agents Group Context-Aware Computing Group MIT Media Laboratory MIT Media Laboratory MIT Media Laboratory Cambridge, MA 02139 Cambridge, MA 02139 Cambridge, MA 02139 +1 617 253 5334 +1 617 253 0315 +1 617 253 6968 [email protected] [email protected] [email protected] ABSTRACT delighted us, the text sits unmoved in cold, square boxes on This paper demonstrates a new approach to recognizing and the computer screen. Nass et al.’s study of human- presenting the affect of text. The approach starts with a computer social interaction reveals that people naturally corpus of 400,000 responses to questions about everyday expect their interactions with computers to be social and life in Open Mind Common Sense. This so-called affective, just as with other people! [20],[21]. commonsense knowledge is the basis of a textual affect Sadly though, people have been so conditioned to expect so sensing engine. The engine dynamically analyzes a user’s little from the user interfaces of today that we are not even text and senses broad affective qualities of the story at the bothered by their inability to affectively respond to us like a sentence level. This paper shows how a commonsense friend or family member might do. affect model was constructed and incorporated into Chernov face style feedback in an affectively responsive This shortcoming in current user interfaces hinders email browser called EmpathyBuddy. This experimental progress in the bigger picture too. If software is to system reacts to sentences as they are typed.
    [Show full text]
  • Bootstrapping Commonsense Knowledge
    ASTRID: Bootstrapping Commonsense Knowledge Hans Peter Willems MIND|CONSTRUCT March 2021 Abstract The need for Commonsense Knowledge in the machine becomes more and more apparent, as we try to move forward in the development of Artificial (General) Intelligence. It is becoming evident that this kind of knowledge is paramount in the human cognitive capacity and therefore also crucial for machine intelligence to ever reach any level of performance nearing that of a human brain. However, attaining a sufficient amount, and qualitative level, of human Commonsense Knowledge in the machine, appears to be an ‘AI­hard’ or ‘AI­complete’ problem. How do humans do this? There is a lot to be learned from child development, and although there are AI­projects that try to use a developmental model to ‘grow’ intelligence, there have not been any (relevant) Commonsense Knowledge projects that leveraged the child development paradigm. That is, until now. I present ASTRID (Analysis of Systemic Tagging Results in Intelligent Dynamics), a real­world implementation of a Cognitive Architecture based on human developmental models. The current state of this project is the result of a full decade of research and development. This paper describes the project background, underlying philosophies, objectives and current results and insights. Keywords: Commonsense Knowledge, Unsupervised Transfer Learning, Machine Intelligence, Semantics, Natural Language Processing, Deep Inference. ©2021 Hans Peter Willems ­ First published March 22, 2021 online @ https://www.mindconstruct.com This work is licensed under a Creative Commons Attribution­ShareAlike 4.0 License 2 ASTRID: Bootstrapping Commonsense Knowledge The case for Commonsense Knowledge As early as 1959, John McCarthy argued for the need of Commonsense Knowledge to attain human level Artificial Intelligence (McCarthy, 1959), currently referred to as Artificial General Intelligence (AGI).
    [Show full text]
  • Natural Language Understanding with Commonsense Reasoning
    E.T.S. DE INGENIEROS INFORMÁTICOS UNIVERSIDAD POLITÉCNICA DE MADRID MASTER TESIS MSc IN ARTIFICIAL INTELLIGENCE (MUIA) NATURAL LANGUAGE UNDERSTANDING WITH COMMONSENSE REASONING: APPLICATION TO THE WINOGRAD SCHEMA CHALLENGE AUTHOR: ALFONSO LÓPEZ TORRES SUPERVISOR: MARTÍN MOLINA GONZÁLEZ JUNE, 2016 This is for my children Carla and Alonso, and my wife Véronique Thanks for their unconditional support and patient (also for the coming adventures…) v Acknowledgments: I’d like to thank the advices and help received from Martín. I was very lucky being your student. vi RESUMEN En 1950, Alan Turing propuso un test para evaluar el grado de inteligencia humana que podría presentar una máquina. La idea principal era realmente sencilla: llevar a cabo una charla abierta entre un evaluador y la máquina. Si dicho evaluador era incapaz de discernir si el examinado era una persona o una máquina, podría afirmarse que el test había sido superado. Desde entonces, a lo largo de los últimos 60 años se han presentado numerosas propuestas a través de los cuales se han puesto al descubierto ciertas debilidades del test. Quizás la más importante es el hecho de centrarse en la inteligencia humana, dejando a un lado otros tipos de inteligencia. El test obliga en gran medida a definir en la máquina un comportamiento antropomórfico y de imitación con el único fin de pasar el test. Con el fin de superar estos y otros puntos débiles, Hector Levesque propuso en 2011 un nuevo reto, “The Winograd Schema Challenge”. Un sencillo test basado en Pregunta y Respuesta sobre una frase que describe una situación cotidiana.
    [Show full text]
  • AI/ML Finding a Doing Machine
    Digital Readiness: AI/ML Finding a doing machine Gregg Vesonder Stevens Institute of Technology http://vesonder.com Three Talks • Digital Readiness: AI/ML, The thinking system quest. – Artificial Intelligence and Machine Learning (AI/ML) have had a fascinating evolution from 1950 to the present. This talk sketches the main themes of AI and machine learning, tracing the evolution of the field since its beginning in the 1950s and explaining some of its main concepts. These eras are characterized as “from knowledge is power” to “data is king”. • Digital Readiness: AI/ML, Finding a doing machine. – In the last decade Machine Learning had a remarkable success record. We will review reasons for that success, review the technology, examine areas of need and explore what happened to the rest of AI, GOFAI (Good Old Fashion AI). • Digital Readiness: AI/ML, Common Sense prevails? – Will there be another AI Winter? We will explore some clues to where the current AI/ML may reunite with GOFAI (Good Old Fashioned AI) and hopefully expand the utility of both. This will include extrapolating on the necessary melding of AI with engineering, particularly systems engineering. Roadmap • Systems – Watson – CYC – NELL – Alexa, Siri, Google Home • Technologies – Semantic web – GPUs and CUDA – Back office (Hadoop) – ML Bias • Last week’s questions Winter is Coming? • First Summer: Irrational Exuberance (1948 – 1966) • First Winter (1967 – 1977) • Second Summer: Knowledge is Power (1978 – 1987) • Second Winter (1988 – 2011) • Third Summer (2012 – ?) • Why there might not be a third winter! Henry Kautz – Engelmore Lecture SYSTEMS Winter 2 Systems • Knowledge is power theme • Influence of the web, try to represent all knowledge – Creating a general ontology organizing everything in the world into a hierarchy of categories – Successful deep ontologies: Gene Ontology and CML Chemical Markup Language • Indeed extreme knowledge – CYC and Open CYC – IBM’s Watson Upper ontology of the world Russell and Norvig figure 12.1 Properties of a subject area and how they are related Ferrucci, D., et.al.
    [Show full text]
  • Conceptualization and Visual Knowledge Organization: a Survey of Ontology-Based Solutions
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Repository of the Academy's Library CONCEPTUALIZATION AND VISUAL KNOWLEDGE ORGANIZATION: A SURVEY OF ONTOLOGY-BASED SOLUTIONS A. Benedek1, G. Lajos2 1 Hungarian Academy of Sciences (HUNGARY) 2 WikiNizer (UNITED KINGDOM) Abstract Conceptualization and Visual Knowledge Organization are overlapping research areas of Intellect Augmentation with hot topics at their intersection which are enhanced by proliferating issues of ontology integration. Knowledge organization is more closely related to the content of concepts and to the actual knowledge items than taxonomic structures and ‘ontologies’ understood as “explicit specifications of a conceptualization” (Gruber). Yet, the alignment of the structure and relationship of knowledge items of a domain of interest require similar operations. We reconsider ontology-based approaches to collaborative conceptualization from the point of view of user interaction in the context of augmented visual knowledge organization. After considering several formal ontology based approaches, it is argued that co-evolving visualization of concepts are needed both at the object and meta-level to handle the social aspects of learning and knowledge work in the framework of an Exploratory Epistemology. Current systems tend to separate the conceptual level from the content, and the context of logical understanding from the intent and use of concepts. Besides, they usually consider the unification of individual contributions in a ubiquitous global representation. As opposed to this God's eye view we are exploring new alternatives that aim to provide Morphic-like live, tinkerable, in situ approaches to conceptualization and its visualization and to direct manipulation based ontology authoring capabilities.
    [Show full text]
  • Toward the Use of Upper Level Ontologies
    Toward the use of upper level ontologies for semantically interoperable systems: an emergency management use case Linda Elmhadhbi, Mohamed Hedi Karray, Bernard Archimède To cite this version: Linda Elmhadhbi, Mohamed Hedi Karray, Bernard Archimède. Toward the use of upper level ontolo- gies for semantically interoperable systems: an emergency management use case. 9th Conference on Interoperability for Enterprise Systems and Applications I-ESA2018 Conference, Mar 2018, Berlin, Germany. pp.1-10. hal-02359706 HAL Id: hal-02359706 https://hal.archives-ouvertes.fr/hal-02359706 Submitted on 12 Nov 2019 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Open Archive Toulouse Archive Ouverte (OATAO ) OATAO is an open access repository that collects the wor of some Toulouse researchers and ma es it freely available over the web where possible. This is an author's version published in: http://oatao.univ-toulouse.fr/22794 Official URL : https://doi.org/10.1007/978-3-030-13693-2 To cite this version: Elmhadhbi, Linda and Karray, Mohamed Hedi and Archimède, Bernard Toward the use of upper level ontologies for semantically interoperable systems: an emergency management use case. ( In Press: 2018) In: 9th Conference on Interoperability for Enterprise Systems and Applications I-ESA2018 Conference, 23 March 2018 - 19 March 2018 (Berlin, Germany).
    [Show full text]
  • Reshare: an Operational Ontology Framework for Research Modeling, Combining and Sharing
    ReShare: An Operational Ontology Framework for Research Modeling, Combining and Sharing Mohammad Al Boni Osama AbuOmar Roger L. King Student Member, IEEE Student Member, IEEE Senior Member, IEEE Center for Advanced Vehicular Systems Center for Advanced Vehicular Systems Center for Advanced Vehicular Systems Mississippi State University, USA Mississippi State University, USA Mississippi State University, USA Email: [email protected] Email: [email protected] Email: [email protected] Abstract—Scientists always face difficulties dealing with dis- Research experiments not only contain semantics which jointed information. There is a need for a standardized and robust can be effectively defined and described using regular ontol- way to represent and exchange knowledge. Ontology has been ogy, but also it contains operations (datasets, source code, anal- widely used for this purpose. However, since research involves ysis, experiments, and/or results). Therefore, an operational semantics and operations, we need to conceptualize both of them. ontology needs to be employed instead of the regular descrip- In this article, we propose ReShare to provide a solution for this tive one. Furthermore, to insure robustness and scalability, we problem. Maximizing utilization while preserving the semantics is one of the main challenges when the heterogeneous knowledge need to design our ontology in such a generic way so that is combined. Therefore, operational annotations were designed it can be easily extended without the need for any change in to allow generic object modeling, binding and representation. the core concepts. As a result, we considered the use of a Furthermore, a test bed is developed and preliminary results are software object oriented (OO) model to insure a high level presented to show the usefulness and robustness of our approach.
    [Show full text]
  • Federated Ontology Search Vasco Calais Pedro CMU-LTI-09-010
    Federated Ontology Search Vasco Calais Pedro CMU-LTI-09-010 Language Technologies Institute School of Computer Science Carnegie Mellon University 5000 Forbes Ave. Pittsburgh, PA 15213 www.lti.cs.cmu.edu Thesis Committee: Jaime Carbonell, Chair Eric Nyberg Robert Frederking Eduard Hovy, Information Sciences Institute Submitted in partial fulfillment of the requirements for the degree Doctor of Philosophy In Language and Information Technologies Copyright © 2009 Vasco Calais Pedro For my grandmother, Avó Helena, I am sorry I wasn’t there Abstract An Ontology can be defined as a formal representation of a set of concepts within a domain and the relationships between those concepts. The development of the semantic web initiative is rapidly increasing the number of publicly available ontologies. In such a distributed environment, complex applications often need to handle multiple ontologies in order to provide adequate domain coverage. Surprisingly, there is a lack of adequate frameworks for enabling the use of multiple ontologies transparently while abstracting the particular ontological structures used by that framework. Given that any ontology represents the views of its author or authors, using multiple ontologies requires us to deal with several significant challenges, some stemming from the nature of knowledge itself, such as cases of polysemy or homography, and some stemming from the structures that we choose to represent such knowledge with. The focus of this thesis is to explore a set of techniques that will allow us to overcome some of the challenges found when using multiple ontologies, thus making progress in the creation of a functional information access platform for structured sources.
    [Show full text]
  • Opportunities and Challenges Presented by Wikidata in the Context of Biocuration
    Opportunities and challenges presented by Wikidata in the context of biocuration 2 3 1 Andra Waagmeester , Elvira Mitraka Benjamin M. Good , Sebastian Burgstaller- 2 1 1 1 Micelio, Veltwijklaan 305, 2180 Antwerp, Belgium Muehlbacher , Tim Putman , Andrew Su 3 1 University of Maryland, School of Medicine, 655 West Department of Molecular and Experimental Medicine, The Baltimore Street, Baltimore, MD, 21201 Scripps Research Institute, 10550 North Torrey Pines Road, La Jolla, CA, 92037 Abstract—Wikidata is a world readable and writable is currently the only major Semantic Web resource that knowledge base maintained by the Wikimedia Foundation. It supports open, collaborative editing. Further, through its offers the opportunity to collaboratively construct a fully open association with the Wikipedias, it has thousands of editors access knowledge graph spanning biology, medicine, and all other working to improve its content. If orchestrated effectively, domains of knowledge. To meet this potential, social and this combination of technology and community could produce technical challenges must be overcome most of which are familiar to the biocuration community. These include a knowledge resource of unprecedented scale and value. In community ontology building, high precision information terms of distributing knowledge, its direct integration with the extraction, provenance, and license management. By working Wikipedias can allow its community vetted content to be together with Wikidata now, we can help shape it into a shared with literally millions of consumers in hundreds of trustworthy, unencumbered central node in the Semantic Web of languages. Outside of the Wikipedias, Wikidata’s CC0 license biomedical data. removes all barriers on re-use and redistribution of its contents in other applications.
    [Show full text]