Redalyc.Cognitive Modules of an NLP Knowledge Base for Language
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
A Survey of Top-Level Ontologies to Inform the Ontological Choices for a Foundation Data Model
A survey of Top-Level Ontologies To inform the ontological choices for a Foundation Data Model Version 1 Contents 1 Introduction and Purpose 3 F.13 FrameNet 92 2 Approach and contents 4 F.14 GFO – General Formal Ontology 94 2.1 Collect candidate top-level ontologies 4 F.15 gist 95 2.2 Develop assessment framework 4 F.16 HQDM – High Quality Data Models 97 2.3 Assessment of candidate top-level ontologies F.17 IDEAS – International Defence Enterprise against the framework 5 Architecture Specification 99 2.4 Terminological note 5 F.18 IEC 62541 100 3 Assessment framework – development basis 6 F.19 IEC 63088 100 3.1 General ontological requirements 6 F.20 ISO 12006-3 101 3.2 Overarching ontological architecture F.21 ISO 15926-2 102 framework 8 F.22 KKO: KBpedia Knowledge Ontology 103 4 Ontological commitment overview 11 F.23 KR Ontology – Knowledge Representation 4.1 General choices 11 Ontology 105 4.2 Formal structure – horizontal and vertical 14 F.24 MarineTLO: A Top-Level 4.3 Universal commitments 33 Ontology for the Marine Domain 106 5 Assessment Framework Results 37 F. 25 MIMOSA CCOM – (Common Conceptual 5.1 General choices 37 Object Model) 108 5.2 Formal structure: vertical aspects 38 F.26 OWL – Web Ontology Language 110 5.3 Formal structure: horizontal aspects 42 F.27 ProtOn – PROTo ONtology 111 5.4 Universal commitments 44 F.28 Schema.org 112 6 Summary 46 F.29 SENSUS 113 Appendix A F.30 SKOS 113 Pathway requirements for a Foundation Data F.31 SUMO 115 Model 48 F.32 TMRM/TMDM – Topic Map Reference/Data Appendix B Models 116 ISO IEC 21838-1:2019 -
Chaudron: Extending Dbpedia with Measurement Julien Subercaze
Chaudron: Extending DBpedia with measurement Julien Subercaze To cite this version: Julien Subercaze. Chaudron: Extending DBpedia with measurement. 14th European Semantic Web Conference, Eva Blomqvist, Diana Maynard, Aldo Gangemi, May 2017, Portoroz, Slovenia. hal- 01477214 HAL Id: hal-01477214 https://hal.archives-ouvertes.fr/hal-01477214 Submitted on 27 Feb 2017 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Chaudron: Extending DBpedia with measurement Julien Subercaze1 Univ Lyon, UJM-Saint-Etienne, CNRS Laboratoire Hubert Curien UMR 5516, F-42023, SAINT-ETIENNE, France [email protected] Abstract. Wikipedia is the largest collaborative encyclopedia and is used as the source for DBpedia, a central dataset of the LOD cloud. Wikipedia contains numerous numerical measures on the entities it describes, as per the general character of the data it encompasses. The DBpedia In- formation Extraction Framework transforms semi-structured data from Wikipedia into structured RDF. However this extraction framework of- fers a limited support to handle measurement in Wikipedia. In this paper, we describe the automated process that enables the creation of the Chaudron dataset. We propose an alternative extraction to the tra- ditional mapping creation from Wikipedia dump, by also using the ren- dered HTML to avoid the template transclusion issue. -
Open Mind Common Sense: Knowledge Acquisition from the General Public
Open Mind Common Sense: Knowledge Acquisition from the General Public Push Singh, Grace Lim, Thomas Lin, Erik T. Mueller Travell Perkins, Mark Tompkins, Wan Li Zhu MIT Media Laboratory 20 Ames Street Cambridge, MA 02139 USA {push, glim, tlin, markt, wlz}@mit.edu, [email protected], [email protected] Abstract underpinnings for commonsense reasoning (Shanahan Open Mind Common Sense is a knowledge acquisition 1997), there has been far less work on finding ways to system designed to acquire commonsense knowledge from accumulate the knowledge to do so in practice. The most the general public over the web. We describe and evaluate well-known attempt has been the Cyc project (Lenat 1995) our first fielded system, which enabled the construction of which contains 1.5 million assertions built over 15 years a 400,000 assertion commonsense knowledge base. We at the cost of several tens of millions of dollars. then discuss how our second-generation system addresses Knowledge bases this large require a tremendous effort to weaknesses discovered in the first. The new system engineer. With the exception of Cyc, this problem of scale acquires facts, descriptions, and stories by allowing has made efforts to study and build commonsense participants to construct and fill in natural language knowledge bases nearly non-existent within the artificial templates. It employs word-sense disambiguation and intelligence community. methods of clarifying entered knowledge, analogical inference to provide feedback, and allows participants to validate knowledge and in turn each other. Turning to the general public 1 In this paper we explore a possible solution to this Introduction problem of scale, based on one critical observation: Every We would like to build software agents that can engage in ordinary person has common sense of the kind we want to commonsense reasoning about ordinary human affairs. -
Automatic Affective Feedback in an Email Browser
Automatic Affective Feedback in an Email Browser Hugo Liu Henry Lieberman Ted Selker Software Agents Group Software Agents Group Context-Aware Computing Group MIT Media Laboratory MIT Media Laboratory MIT Media Laboratory Cambridge, MA 02139 Cambridge, MA 02139 Cambridge, MA 02139 +1 617 253 5334 +1 617 253 0315 +1 617 253 6968 [email protected] [email protected] [email protected] ABSTRACT delighted us, the text sits unmoved in cold, square boxes on This paper demonstrates a new approach to recognizing and the computer screen. Nass et al.’s study of human- presenting the affect of text. The approach starts with a computer social interaction reveals that people naturally corpus of 400,000 responses to questions about everyday expect their interactions with computers to be social and life in Open Mind Common Sense. This so-called affective, just as with other people! [20],[21]. commonsense knowledge is the basis of a textual affect Sadly though, people have been so conditioned to expect so sensing engine. The engine dynamically analyzes a user’s little from the user interfaces of today that we are not even text and senses broad affective qualities of the story at the bothered by their inability to affectively respond to us like a sentence level. This paper shows how a commonsense friend or family member might do. affect model was constructed and incorporated into Chernov face style feedback in an affectively responsive This shortcoming in current user interfaces hinders email browser called EmpathyBuddy. This experimental progress in the bigger picture too. If software is to system reacts to sentences as they are typed. -
Bootstrapping Commonsense Knowledge
ASTRID: Bootstrapping Commonsense Knowledge Hans Peter Willems MIND|CONSTRUCT March 2021 Abstract The need for Commonsense Knowledge in the machine becomes more and more apparent, as we try to move forward in the development of Artificial (General) Intelligence. It is becoming evident that this kind of knowledge is paramount in the human cognitive capacity and therefore also crucial for machine intelligence to ever reach any level of performance nearing that of a human brain. However, attaining a sufficient amount, and qualitative level, of human Commonsense Knowledge in the machine, appears to be an ‘AIhard’ or ‘AIcomplete’ problem. How do humans do this? There is a lot to be learned from child development, and although there are AIprojects that try to use a developmental model to ‘grow’ intelligence, there have not been any (relevant) Commonsense Knowledge projects that leveraged the child development paradigm. That is, until now. I present ASTRID (Analysis of Systemic Tagging Results in Intelligent Dynamics), a realworld implementation of a Cognitive Architecture based on human developmental models. The current state of this project is the result of a full decade of research and development. This paper describes the project background, underlying philosophies, objectives and current results and insights. Keywords: Commonsense Knowledge, Unsupervised Transfer Learning, Machine Intelligence, Semantics, Natural Language Processing, Deep Inference. ©2021 Hans Peter Willems First published March 22, 2021 online @ https://www.mindconstruct.com This work is licensed under a Creative Commons AttributionShareAlike 4.0 License 2 ASTRID: Bootstrapping Commonsense Knowledge The case for Commonsense Knowledge As early as 1959, John McCarthy argued for the need of Commonsense Knowledge to attain human level Artificial Intelligence (McCarthy, 1959), currently referred to as Artificial General Intelligence (AGI). -
Natural Language Understanding with Commonsense Reasoning
E.T.S. DE INGENIEROS INFORMÁTICOS UNIVERSIDAD POLITÉCNICA DE MADRID MASTER TESIS MSc IN ARTIFICIAL INTELLIGENCE (MUIA) NATURAL LANGUAGE UNDERSTANDING WITH COMMONSENSE REASONING: APPLICATION TO THE WINOGRAD SCHEMA CHALLENGE AUTHOR: ALFONSO LÓPEZ TORRES SUPERVISOR: MARTÍN MOLINA GONZÁLEZ JUNE, 2016 This is for my children Carla and Alonso, and my wife Véronique Thanks for their unconditional support and patient (also for the coming adventures…) v Acknowledgments: I’d like to thank the advices and help received from Martín. I was very lucky being your student. vi RESUMEN En 1950, Alan Turing propuso un test para evaluar el grado de inteligencia humana que podría presentar una máquina. La idea principal era realmente sencilla: llevar a cabo una charla abierta entre un evaluador y la máquina. Si dicho evaluador era incapaz de discernir si el examinado era una persona o una máquina, podría afirmarse que el test había sido superado. Desde entonces, a lo largo de los últimos 60 años se han presentado numerosas propuestas a través de los cuales se han puesto al descubierto ciertas debilidades del test. Quizás la más importante es el hecho de centrarse en la inteligencia humana, dejando a un lado otros tipos de inteligencia. El test obliga en gran medida a definir en la máquina un comportamiento antropomórfico y de imitación con el único fin de pasar el test. Con el fin de superar estos y otros puntos débiles, Hector Levesque propuso en 2011 un nuevo reto, “The Winograd Schema Challenge”. Un sencillo test basado en Pregunta y Respuesta sobre una frase que describe una situación cotidiana. -
AI/ML Finding a Doing Machine
Digital Readiness: AI/ML Finding a doing machine Gregg Vesonder Stevens Institute of Technology http://vesonder.com Three Talks • Digital Readiness: AI/ML, The thinking system quest. – Artificial Intelligence and Machine Learning (AI/ML) have had a fascinating evolution from 1950 to the present. This talk sketches the main themes of AI and machine learning, tracing the evolution of the field since its beginning in the 1950s and explaining some of its main concepts. These eras are characterized as “from knowledge is power” to “data is king”. • Digital Readiness: AI/ML, Finding a doing machine. – In the last decade Machine Learning had a remarkable success record. We will review reasons for that success, review the technology, examine areas of need and explore what happened to the rest of AI, GOFAI (Good Old Fashion AI). • Digital Readiness: AI/ML, Common Sense prevails? – Will there be another AI Winter? We will explore some clues to where the current AI/ML may reunite with GOFAI (Good Old Fashioned AI) and hopefully expand the utility of both. This will include extrapolating on the necessary melding of AI with engineering, particularly systems engineering. Roadmap • Systems – Watson – CYC – NELL – Alexa, Siri, Google Home • Technologies – Semantic web – GPUs and CUDA – Back office (Hadoop) – ML Bias • Last week’s questions Winter is Coming? • First Summer: Irrational Exuberance (1948 – 1966) • First Winter (1967 – 1977) • Second Summer: Knowledge is Power (1978 – 1987) • Second Winter (1988 – 2011) • Third Summer (2012 – ?) • Why there might not be a third winter! Henry Kautz – Engelmore Lecture SYSTEMS Winter 2 Systems • Knowledge is power theme • Influence of the web, try to represent all knowledge – Creating a general ontology organizing everything in the world into a hierarchy of categories – Successful deep ontologies: Gene Ontology and CML Chemical Markup Language • Indeed extreme knowledge – CYC and Open CYC – IBM’s Watson Upper ontology of the world Russell and Norvig figure 12.1 Properties of a subject area and how they are related Ferrucci, D., et.al. -
Conceptualization and Visual Knowledge Organization: a Survey of Ontology-Based Solutions
View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Repository of the Academy's Library CONCEPTUALIZATION AND VISUAL KNOWLEDGE ORGANIZATION: A SURVEY OF ONTOLOGY-BASED SOLUTIONS A. Benedek1, G. Lajos2 1 Hungarian Academy of Sciences (HUNGARY) 2 WikiNizer (UNITED KINGDOM) Abstract Conceptualization and Visual Knowledge Organization are overlapping research areas of Intellect Augmentation with hot topics at their intersection which are enhanced by proliferating issues of ontology integration. Knowledge organization is more closely related to the content of concepts and to the actual knowledge items than taxonomic structures and ‘ontologies’ understood as “explicit specifications of a conceptualization” (Gruber). Yet, the alignment of the structure and relationship of knowledge items of a domain of interest require similar operations. We reconsider ontology-based approaches to collaborative conceptualization from the point of view of user interaction in the context of augmented visual knowledge organization. After considering several formal ontology based approaches, it is argued that co-evolving visualization of concepts are needed both at the object and meta-level to handle the social aspects of learning and knowledge work in the framework of an Exploratory Epistemology. Current systems tend to separate the conceptual level from the content, and the context of logical understanding from the intent and use of concepts. Besides, they usually consider the unification of individual contributions in a ubiquitous global representation. As opposed to this God's eye view we are exploring new alternatives that aim to provide Morphic-like live, tinkerable, in situ approaches to conceptualization and its visualization and to direct manipulation based ontology authoring capabilities. -
Toward the Use of Upper Level Ontologies
Toward the use of upper level ontologies for semantically interoperable systems: an emergency management use case Linda Elmhadhbi, Mohamed Hedi Karray, Bernard Archimède To cite this version: Linda Elmhadhbi, Mohamed Hedi Karray, Bernard Archimède. Toward the use of upper level ontolo- gies for semantically interoperable systems: an emergency management use case. 9th Conference on Interoperability for Enterprise Systems and Applications I-ESA2018 Conference, Mar 2018, Berlin, Germany. pp.1-10. hal-02359706 HAL Id: hal-02359706 https://hal.archives-ouvertes.fr/hal-02359706 Submitted on 12 Nov 2019 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Open Archive Toulouse Archive Ouverte (OATAO ) OATAO is an open access repository that collects the wor of some Toulouse researchers and ma es it freely available over the web where possible. This is an author's version published in: http://oatao.univ-toulouse.fr/22794 Official URL : https://doi.org/10.1007/978-3-030-13693-2 To cite this version: Elmhadhbi, Linda and Karray, Mohamed Hedi and Archimède, Bernard Toward the use of upper level ontologies for semantically interoperable systems: an emergency management use case. ( In Press: 2018) In: 9th Conference on Interoperability for Enterprise Systems and Applications I-ESA2018 Conference, 23 March 2018 - 19 March 2018 (Berlin, Germany). -
Reshare: an Operational Ontology Framework for Research Modeling, Combining and Sharing
ReShare: An Operational Ontology Framework for Research Modeling, Combining and Sharing Mohammad Al Boni Osama AbuOmar Roger L. King Student Member, IEEE Student Member, IEEE Senior Member, IEEE Center for Advanced Vehicular Systems Center for Advanced Vehicular Systems Center for Advanced Vehicular Systems Mississippi State University, USA Mississippi State University, USA Mississippi State University, USA Email: [email protected] Email: [email protected] Email: [email protected] Abstract—Scientists always face difficulties dealing with dis- Research experiments not only contain semantics which jointed information. There is a need for a standardized and robust can be effectively defined and described using regular ontol- way to represent and exchange knowledge. Ontology has been ogy, but also it contains operations (datasets, source code, anal- widely used for this purpose. However, since research involves ysis, experiments, and/or results). Therefore, an operational semantics and operations, we need to conceptualize both of them. ontology needs to be employed instead of the regular descrip- In this article, we propose ReShare to provide a solution for this tive one. Furthermore, to insure robustness and scalability, we problem. Maximizing utilization while preserving the semantics is one of the main challenges when the heterogeneous knowledge need to design our ontology in such a generic way so that is combined. Therefore, operational annotations were designed it can be easily extended without the need for any change in to allow generic object modeling, binding and representation. the core concepts. As a result, we considered the use of a Furthermore, a test bed is developed and preliminary results are software object oriented (OO) model to insure a high level presented to show the usefulness and robustness of our approach. -
Federated Ontology Search Vasco Calais Pedro CMU-LTI-09-010
Federated Ontology Search Vasco Calais Pedro CMU-LTI-09-010 Language Technologies Institute School of Computer Science Carnegie Mellon University 5000 Forbes Ave. Pittsburgh, PA 15213 www.lti.cs.cmu.edu Thesis Committee: Jaime Carbonell, Chair Eric Nyberg Robert Frederking Eduard Hovy, Information Sciences Institute Submitted in partial fulfillment of the requirements for the degree Doctor of Philosophy In Language and Information Technologies Copyright © 2009 Vasco Calais Pedro For my grandmother, Avó Helena, I am sorry I wasn’t there Abstract An Ontology can be defined as a formal representation of a set of concepts within a domain and the relationships between those concepts. The development of the semantic web initiative is rapidly increasing the number of publicly available ontologies. In such a distributed environment, complex applications often need to handle multiple ontologies in order to provide adequate domain coverage. Surprisingly, there is a lack of adequate frameworks for enabling the use of multiple ontologies transparently while abstracting the particular ontological structures used by that framework. Given that any ontology represents the views of its author or authors, using multiple ontologies requires us to deal with several significant challenges, some stemming from the nature of knowledge itself, such as cases of polysemy or homography, and some stemming from the structures that we choose to represent such knowledge with. The focus of this thesis is to explore a set of techniques that will allow us to overcome some of the challenges found when using multiple ontologies, thus making progress in the creation of a functional information access platform for structured sources. -
Opportunities and Challenges Presented by Wikidata in the Context of Biocuration
Opportunities and challenges presented by Wikidata in the context of biocuration 2 3 1 Andra Waagmeester , Elvira Mitraka Benjamin M. Good , Sebastian Burgstaller- 2 1 1 1 Micelio, Veltwijklaan 305, 2180 Antwerp, Belgium Muehlbacher , Tim Putman , Andrew Su 3 1 University of Maryland, School of Medicine, 655 West Department of Molecular and Experimental Medicine, The Baltimore Street, Baltimore, MD, 21201 Scripps Research Institute, 10550 North Torrey Pines Road, La Jolla, CA, 92037 Abstract—Wikidata is a world readable and writable is currently the only major Semantic Web resource that knowledge base maintained by the Wikimedia Foundation. It supports open, collaborative editing. Further, through its offers the opportunity to collaboratively construct a fully open association with the Wikipedias, it has thousands of editors access knowledge graph spanning biology, medicine, and all other working to improve its content. If orchestrated effectively, domains of knowledge. To meet this potential, social and this combination of technology and community could produce technical challenges must be overcome most of which are familiar to the biocuration community. These include a knowledge resource of unprecedented scale and value. In community ontology building, high precision information terms of distributing knowledge, its direct integration with the extraction, provenance, and license management. By working Wikipedias can allow its community vetted content to be together with Wikidata now, we can help shape it into a shared with literally millions of consumers in hundreds of trustworthy, unencumbered central node in the Semantic Web of languages. Outside of the Wikipedias, Wikidata’s CC0 license biomedical data. removes all barriers on re-use and redistribution of its contents in other applications.