Semantic Networks
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
1992-Common Sense Retrieval
From: AAAI-92 Proceedings. Copyright ©1992, AAAI (www.aaai.org). All rights reserved. Common Sense A. Julian Graddock University of B.C. Vancouver, B.C., V6T lW5 [email protected] Abstract Such inferences are made by retrieving common sense knowledge: General stable knowledge about the An important and readily available source unobserved properties obtained by aggregating over a of knowledge for common sense reasoning is set of similar experiences in the EKB. Thus, the CSR partial descriptions of specific experiences. might infer that the bird sitting on the lawn flys if Knowledge bases (KBs) containing such in- flying is part of the common sense knowledge about formation are called episodic knowledge buses similar birds. (EKB). Aggregations of episodic knowledge In this document I provide a first order framework provide common sense knowledge about the appropriate for: 1. Describing experiences, and 2. Mak- unobserved properties of ‘new’ experiences. ing inferences by retrieving common sense knowledge Such knowledge is retrieved by applying directly from an EKB. The framework differs from statistics to a relevant subset of the EKB many existing knowledge representation techniques as called the reference class. it provides information about specific experiences. I I study a manner in which a corpus of expe- judge the effectiveness of my model by its ability to riences can be represented to allow common flexibly and efficiently retrieve common sense knowl- sense retrieval which is: 1. Flexible enough edge. to allow the common sense reasoner to deal with ‘new’ experiences, and 2. In the simplest My solution case, reduces to efficient database look-up. -
University of Navarra Ecclesiastical Faculty of Philosophy
UNIVERSITY OF NAVARRA ECCLESIASTICAL FACULTY OF PHILOSOPHY Mark Telford Georges THE PROBLEM OF STORING COMMON SENSE IN ARTIFICIAL INTELLIGENCE Context in CYC Doctoral dissertation directed by Dr. Jaime Nubiola Pamplona, 2000 1 Table of Contents ABBREVIATIONS ............................................................................................................................. 4 INTRODUCTION............................................................................................................................... 5 CHAPTER I: A SHORT HISTORY OF ARTIFICIAL INTELLIGENCE .................................. 9 1.1. THE ORIGIN AND USE OF THE TERM “ARTIFICIAL INTELLIGENCE”.............................................. 9 1.1.1. Influences in AI................................................................................................................ 10 1.1.2. “Artificial Intelligence” in popular culture..................................................................... 11 1.1.3. “Artificial Intelligence” in Applied AI ............................................................................ 12 1.1.4. Human AI and alien AI....................................................................................................14 1.1.5. “Artificial Intelligence” in Cognitive Science................................................................. 16 1.2. TRENDS IN AI........................................................................................................................... 17 1.2.1. Classical AI .................................................................................................................... -
Probabilistic Topic Modelling with Semantic Graph
Probabilistic Topic Modelling with Semantic Graph B Long Chen( ), Joemon M. Jose, Haitao Yu, Fajie Yuan, and Huaizhi Zhang School of Computing Science, University of Glasgow, Sir Alwyns Building, Glasgow, UK [email protected] Abstract. In this paper we propose a novel framework, topic model with semantic graph (TMSG), which couples topic model with the rich knowledge from DBpedia. To begin with, we extract the disambiguated entities from the document collection using a document entity linking system, i.e., DBpedia Spotlight, from which two types of entity graphs are created from DBpedia to capture local and global contextual knowl- edge, respectively. Given the semantic graph representation of the docu- ments, we propagate the inherent topic-document distribution with the disambiguated entities of the semantic graphs. Experiments conducted on two real-world datasets show that TMSG can significantly outperform the state-of-the-art techniques, namely, author-topic Model (ATM) and topic model with biased propagation (TMBP). Keywords: Topic model · Semantic graph · DBpedia 1 Introduction Topic models, such as Probabilistic Latent Semantic Analysis (PLSA) [7]and Latent Dirichlet Analysis (LDA) [2], have been remarkably successful in ana- lyzing textual content. Specifically, each document in a document collection is represented as random mixtures over latent topics, where each topic is character- ized by a distribution over words. Such a paradigm is widely applied in various areas of text mining. In view of the fact that the information used by these mod- els are limited to document collection itself, some recent progress have been made on incorporating external resources, such as time [8], geographic location [12], and authorship [15], into topic models. -
Semantic Memory: a Review of Methods, Models, and Current Challenges
Psychonomic Bulletin & Review https://doi.org/10.3758/s13423-020-01792-x Semantic memory: A review of methods, models, and current challenges Abhilasha A. Kumar1 # The Psychonomic Society, Inc. 2020 Abstract Adult semantic memory has been traditionally conceptualized as a relatively static memory system that consists of knowledge about the world, concepts, and symbols. Considerable work in the past few decades has challenged this static view of semantic memory, and instead proposed a more fluid and flexible system that is sensitive to context, task demands, and perceptual and sensorimotor information from the environment. This paper (1) reviews traditional and modern computational models of seman- tic memory, within the umbrella of network (free association-based), feature (property generation norms-based), and distribu- tional semantic (natural language corpora-based) models, (2) discusses the contribution of these models to important debates in the literature regarding knowledge representation (localist vs. distributed representations) and learning (error-free/Hebbian learning vs. error-driven/predictive learning), and (3) evaluates how modern computational models (neural network, retrieval- based, and topic models) are revisiting the traditional “static” conceptualization of semantic memory and tackling important challenges in semantic modeling such as addressing temporal, contextual, and attentional influences, as well as incorporating grounding and compositionality into semantic representations. The review also identifies new challenges -
Knowledge Graphs on the Web – an Overview Arxiv:2003.00719V3 [Cs
January 2020 Knowledge Graphs on the Web – an Overview Nicolas HEIST, Sven HERTLING, Daniel RINGLER, and Heiko PAULHEIM Data and Web Science Group, University of Mannheim, Germany Abstract. Knowledge Graphs are an emerging form of knowledge representation. While Google coined the term Knowledge Graph first and promoted it as a means to improve their search results, they are used in many applications today. In a knowl- edge graph, entities in the real world and/or a business domain (e.g., people, places, or events) are represented as nodes, which are connected by edges representing the relations between those entities. While companies such as Google, Microsoft, and Facebook have their own, non-public knowledge graphs, there is also a larger body of publicly available knowledge graphs, such as DBpedia or Wikidata. In this chap- ter, we provide an overview and comparison of those publicly available knowledge graphs, and give insights into their contents, size, coverage, and overlap. Keywords. Knowledge Graph, Linked Data, Semantic Web, Profiling 1. Introduction Knowledge Graphs are increasingly used as means to represent knowledge. Due to their versatile means of representation, they can be used to integrate different heterogeneous data sources, both within as well as across organizations. [8,9] Besides such domain-specific knowledge graphs which are typically developed for specific domains and/or use cases, there are also public, cross-domain knowledge graphs encoding common knowledge, such as DBpedia, Wikidata, or YAGO. [33] Such knowl- edge graphs may be used, e.g., for automatically enriching data with background knowl- arXiv:2003.00719v3 [cs.AI] 12 Mar 2020 edge to be used in knowledge-intensive downstream applications. -
Large Semantic Network Manual Annotation 1 Introduction
Large Semantic Network Manual Annotation V´aclav Nov´ak Institute of Formal and Applied Linguistics Charles University, Prague [email protected] Abstract This abstract describes a project aiming at manual annotation of the content of natural language utterances in a parallel text corpora. The formalism used in this project is MultiNet – Multilayered Ex- tended Semantic Network. The annotation should be incorporated into Prague Dependency Treebank as a new annotation layer. 1 Introduction A formal specification of the semantic content is the aim of numerous semantic approaches such as TIL [6], DRT [9], MultiNet [4], and others. As far as we can tell, there is no large “real life” text corpora manually annotated with such markup. The projects usually work only with automatically generated annotation, if any [1, 6, 3, 2]. We want to create a parallel Czech-English corpora of texts annotated with the corresponding semantic network. 1.1 Prague Dependency Treebank From the linguistic viewpoint there language resources such as Prague Dependency Treebank (PDT) which contain a deep manual analysis of texts [8]. PDT contains annotations of three layers, namely morpho- logical, analytical (shallow dependency syntax) and tectogrammatical (deep dependency syntax). The units of each annotation level are linked with corresponding units on the preceding level. The morpho- logical units are linked directly with the original text. The theoretical basis of the treebank lies in the Functional Gener- ative Description of language system [7]. PDT 2.0 is based on the long-standing Praguian linguistic tradi- tion, adapted for the current computational-linguistics research needs. The corpus itself is embedded into the latest annotation technology. -
Universal Or Variation? Semantic Networks in English and Chinese
Universal or variation? Semantic networks in English and Chinese Understanding the structures of semantic networks can provide great insights into lexico- semantic knowledge representation. Previous work reveals small-world structure in English, the structure that has the following properties: short average path lengths between words and strong local clustering, with a scale-free distribution in which most nodes have few connections while a small number of nodes have many connections1. However, it is not clear whether such semantic network properties hold across human languages. In this study, we investigate the universal structures and cross-linguistic variations by comparing the semantic networks in English and Chinese. Network description To construct the Chinese and the English semantic networks, we used Chinese Open Wordnet2,3 and English WordNet4. The two wordnets have different word forms in Chinese and English but common word meanings. Word meanings are connected not only to word forms, but also to other word meanings if they form relations such as hypernyms and meronyms (Figure 1). 1. Cross-linguistic comparisons Analysis The large-scale structures of the Chinese and the English networks were measured with two key network metrics, small-worldness5 and scale-free distribution6. Results The two networks have similar size and both exhibit small-worldness (Table 1). However, the small-worldness is much greater in the Chinese network (σ = 213.35) than in the English network (σ = 83.15); this difference is primarily due to the higher average clustering coefficient (ACC) of the Chinese network. The scale-free distributions are similar across the two networks, as indicated by ANCOVA, F (1, 48) = 0.84, p = .37. -
Udc 004.838.2 Irsti 20.19.21 Comparative Analysis Of
Comparative analysis of knowledge bases: ConceptNet vs CYC A. Y. Nuraliyeva, S. S. Daukishov, R.T. Nassyrova UDC 004.838.2 IRSTI 20.19.21 COMPARATIVE ANALYSIS OF KNOWLEDGE BASES: CONCEPTNET VS CYC A. Y. Nuraliyeva1, S. S. Daukishov2, R.T. Nassyrova3 1,2Kazakh-British Technical University, Almaty, Kazakhstan 3IT Analyst, Philip Morris Kazakhstan, Almaty, Kazakhstan [email protected], ORCID 0000-0001-6451-3743 [email protected], ORCID 0000-0002-3784-4456 [email protected], ORCID 0000-0002-7968-3195 Abstract. Data summarization, question answering, text categorization are some of the tasks knowledge bases are used for. A knowledge base (KB) is a computerized compilation of information about the world. They can be useful for complex tasks and problems in NLP and they comprise both entities and relations. We performed an analytical comparison of two knowledge bases which capture a wide variety of common-sense information - ConceptNet and Cyc. Manually curated knowledge base Cyc has invested more than 1,000 man-years over the last two decades building a knowledge base that is meant to capture a wide range of common-sense skills. On the other hand, ConceptNet is a free multilingual knowledge base and crowdsourced knowledge initiative that uses a large number of links to connect commonplace items. When looking for common sense logic and answers to questions, ConceptNet is a great place to start. In this research, two well-known knowledge bases were reviewed - ConceptNet and Cyc - their origin, differences, applications, benefits and disadvantages were covered. The authors hope this paper would be useful for researchers looking for more appropriate knowledge bases for word embedding, word sense disambiguation and natural-language communication. -
Social Engineering in Cybersecurity: a Domain Ontology and Knowledge Graph Application Examples Zuoguang Wang1,2* , Hongsong Zhu1,2* ,Peipeiliu1,2 and Limin Sun1,2
Wang et al. Cybersecurity (2021) 4:31 Cybersecurity https://doi.org/10.1186/s42400-021-00094-6 RESEARCH Open Access Social engineering in cybersecurity: a domain ontology and knowledge graph application examples Zuoguang Wang1,2* , Hongsong Zhu1,2* ,PeipeiLiu1,2 and Limin Sun1,2 Abstract Social engineering has posed a serious threat to cyberspace security. To protect against social engineering attacks, a fundamental work is to know what constitutes social engineering. This paper first develops a domain ontology of social engineering in cybersecurity and conducts ontology evaluation by its knowledge graph application. The domain ontology defines 11 concepts of core entities that significantly constitute or affect social engineering domain, together with 22 kinds of relations describing how these entities related to each other. It provides a formal and explicit knowledge schema to understand, analyze, reuse and share domain knowledge of social engineering. Furthermore, this paper builds a knowledge graph based on 15 social engineering attack incidents and scenarios. 7 knowledge graph application examples (in 6 analysis patterns) demonstrate that the ontology together with knowledge graph is useful to 1) understand and analyze social engineering attack scenario and incident, 2) find the top ranked social engineering threat elements (e.g. the most exploited human vulnerabilities and most used attack mediums), 3) find potential social engineering threats to victims, 4) find potential targets for social engineering attackers, 5) find potential attack paths from specific attacker to specific target, and 6) analyze the same origin attacks. Keywords: Social engineering attack, Cyber security, Ontology, Knowledge graph, Attack scenarios, Threat analysis, Attack path, Attack model, Taxonomy, Composition and structure Introduction (Wang et al. -
November 15, 2019, NIH Record, Vol. LXXI, No. 23
November 15, 2019 Vol. LXXI, No. 23 with a brain-controlled robotic exoskeleton who had been paralyzed for 9 years, “didn’t say, ‘I kicked the ball!’” said Nicolelis. More CLINICAL PROMISE SHOWN importantly, he said, “I felt the ball!” It took a team of 156 scientists from Nicolelis Outlines Progress in 25 countries on 5 continents to reach this Brain-Machine Interfaces moment for which neuroscientist Nicolelis BY RICH MCMANUS had been preparing for 20 years. He recruited the team by dangling field tickets to There is probably no other scientist in the World Cup in front of potential recruits. the world for whom peer review meant “It was a hard way to win a free ticket having his experiment succeed in front of a to the game, but that is the Brazilian way,” stadium full of 75,000 screaming Brazilians, quipped Nicolelis, a native of that country. with another 1.2 billion people watching on Now a professor of neuroscience at live television. Duke University School of Medicine, But at the start of the 2014 World Cup at Nicolelis, who won an NIH Pioneer Award Corinthians Arena in Sao Paulo, Dr. Miguel in 2010 for work he said couldn’t earn a Nicolelis witnessed his patient Juliano Pinto, penny of funding a decade earlier, spoke a paraplegic, not only kick a soccer ball to Oct. 16 at an NIH Director’s Lecture in start the tournament, but also “feel” his foot Masur Auditorium. striking the ball. In the late 1980s, Nicolelis, who has been Duke’s Dr. Miguel Nicolelis discusses his research on brain-machine interfaces. -
Vivid Knowledge and Tractable Reasoning: Preliminary Report
Vivid Knowledge and Tractable Reasoning: Preliminary Report David W. Etherington Alex Borgida Ronald J. Brachman AT&T Bell Laboratories Department of Computer Science Henry Kautz 600 Mountain Avenue Rutgers University AT&T Bell Laboratories Murray Hill, NJ 07974-2070 New Brunswick, NJ Murray Hill, NJ 07974-2070 Abstract can be expressed in the knowledge base (KB) is restricted (sometimes severely) to guarantee that queries can be Mundane, everyday, reasoning is fast. Given answered in more or less reasonable time. In the lat• the inherent complexity of sound and com• ter, restrictions like avoiding chaining or four-valued in• plete reasoning with representations expressive terpretations yield limited conclusions, albeit from rela• enough to capture what people seem to know, tively expressive KB's. commonsense reasoning must require shortcuts We conjecture that a key to efficient problem-solving and assumptions. Some means of simplifying lies in a notion of commonsense reasoning—the kind of the retrieval of the inferential consequences of reasoning that people engage in all the time without re• a set of facts is obviously required. Instead of course to "paper and pencil'1, reasoning by cases, back• looking, as others have, at limited inference or tracking, or particularly deep thought.1 Commonsense syntactic restrictions on the representation, we reasoning is fast: if it were a problem-solver's normal explore the use of "vivid" forms for knowledge, mode of reasoning, then the problem-solver would be in which determining the truth of a sentence is fast. Paradoxically, studies of commonsense reasoning on the order of a database retrieval. in AI (e.g., nonmonotonic logics) have frequently led to In order to base a reasoning system on mechanisms that are even less tractable than logical de• vivid knowledge, we consider ways to construct duction. -
Using a Natural Language Understanding System to Generate Semantic Web Content
Final version to appear, International Journal on Semantic Web and Information Systems, 3(4), 2007. (http://www.igi-pub.com/) 1 Using a Natural Language Understanding System to Generate Semantic Web Content Akshay Java, Sergei Nirneburg, Marjorie McShane, Timothy Finin, Jesse English, Anupam Joshi University of Maryland, Baltimore County, Baltimore MD 21250 {aks1,sergei,marge,finin,english1,joshi}@umbc.edu Abstract We describe our research on automatically generating rich semantic annotations of text and making it available on the Semantic Web. In particular, we discuss the challenges involved in adapting the OntoSem natural language processing system for this purpose. OntoSem, an implementation of the theory of ontological semantics under continuous development for over fifteen years, uses a specially constructed NLP-oriented ontology and an ontological- semantic lexicon to translate English text into a custom ontology-motivated knowledge representation language, the language of text meaning representations (TMRs). OntoSem concentrates on a variety of ambiguity resolution tasks as well as processing unexpected input and reference. To adapt OntoSem’s representation to the Semantic Web, we developed a translation system, OntoSem2OWL, between the TMR language into the Semantic Web language OWL. We next used OntoSem and OntoSem2OWL to support SemNews, an experimental web service that monitors RSS news sources, processes the summaries of the news stories and publishes a structured representation of the meaning of the text in the news story. Index Terms semantic web, OWL, RDF, natural language processing, information extraction I. INTRODUCTION A core goal of the development of the Semantic Web is to bring progressively more meaning to the information published on the Web.