AI/ML Finding a Doing Machine

Total Page:16

File Type:pdf, Size:1020Kb

AI/ML Finding a Doing Machine Digital Readiness: AI/ML Finding a doing machine Gregg Vesonder Stevens Institute of Technology http://vesonder.com Three Talks • Digital Readiness: AI/ML, The thinking system quest. – Artificial Intelligence and Machine Learning (AI/ML) have had a fascinating evolution from 1950 to the present. This talk sketches the main themes of AI and machine learning, tracing the evolution of the field since its beginning in the 1950s and explaining some of its main concepts. These eras are characterized as “from knowledge is power” to “data is king”. • Digital Readiness: AI/ML, Finding a doing machine. – In the last decade Machine Learning had a remarkable success record. We will review reasons for that success, review the technology, examine areas of need and explore what happened to the rest of AI, GOFAI (Good Old Fashion AI). • Digital Readiness: AI/ML, Common Sense prevails? – Will there be another AI Winter? We will explore some clues to where the current AI/ML may reunite with GOFAI (Good Old Fashioned AI) and hopefully expand the utility of both. This will include extrapolating on the necessary melding of AI with engineering, particularly systems engineering. Roadmap • Systems – Watson – CYC – NELL – Alexa, Siri, Google Home • Technologies – Semantic web – GPUs and CUDA – Back office (Hadoop) – ML Bias • Last week’s questions Winter is Coming? • First Summer: Irrational Exuberance (1948 – 1966) • First Winter (1967 – 1977) • Second Summer: Knowledge is Power (1978 – 1987) • Second Winter (1988 – 2011) • Third Summer (2012 – ?) • Why there might not be a third winter! Henry Kautz – Engelmore Lecture SYSTEMS Winter 2 Systems • Knowledge is power theme • Influence of the web, try to represent all knowledge – Creating a general ontology organizing everything in the world into a hierarchy of categories – Successful deep ontologies: Gene Ontology and CML Chemical Markup Language • Indeed extreme knowledge – CYC and Open CYC – IBM’s Watson Upper ontology of the world Russell and Norvig figure 12.1 Properties of a subject area and how they are related Ferrucci, D., et.al. “Building Watson: An overview of the DeepQA Project”, AI Magazine, Fall 2010. Watson • IBM Watson question answering computer system – DeepQA project • Famous for winning a Jeopardy contest against Ken Jennings (KenJen) and Brad Rudder, 2 champions – Jeopardy is difficult, ambiguous questions and average of 3 seconds to ring in – Excluded A/V questions and special instruction questions • Applications: tax preparation, teaching assistant, medicine, … • “ … Create general purpose, reusable natural language processing (NLP) and knowledge representation and reasoning (KRR) that can exploit as-is natural language resources and as- is structured knowledge rather than curate task-specific knowledge resources.” Distribution of Jeopardy Questions KenJen TREC Text Retrieval Contest https://trec.nist.gov/ • to encourage research in information retrieval based on large test collections; • to increase communication among industry, academia, and government by creating an open forum for the exchange of research ideas; • to speed the transfer of technology from research labs into commercial products by demonstrating substantial improvements in retrieval methodologies on real-world problems; and • to increase the availability of appropriate evaluation techniques for use by industry and academia, including development of new evaluation techniques more applicable to current systems. DeepQA Technology • Used more than 100 different techniques • Not specific to Jeopardy but specific to QA • Technological Principles involved: – Massive parallelism – Many Experts – Pervasive confidence estimates – Integrate shallow and deep knowledge – leveraging many loosely formed ontologies – No one algorithm dominates DeepQA High-Level Architecture DeepQA - wikipedia Deep QA/Watson • Uses Hadoop for distributed computing • Watson can process 500 gigabytes/sec, equivalent of 1 million books/sec • Resources included taxonomies and ontologies including DBpedia, WordNet and YAGO CYC • Goal to assemble a comprehensive ontology and knowledge base that spans the basic concepts and rules about how the world works hoping to capture common sense and implicit/tacit knowledge. • Combat brittle AI systems • Began in 1984 by Doug Lenat • Also OpenCyc (CYC) and ResearchCYC (Cyc) Thanks Wikipedia CYC factoids • Ontology (2017) 1.5 Million terms: – 416K collections e.g., fish and types of fishing • CYC KB of common-sense rules and assertions were largely hand-written 24.5 million taking >1K person years! • Uses community of agents (blackboard) architecture • ”every tree is a plant”, “plants die eventually”, asked whether trees die it will infer correctly CYC Elements • Individual items #$BillClinton or #$France (CycL) • Collections #$Tree-ThePlant – contains all trees • Functions produce new terms from given ones. #$FruitFn arg will return collection of fruits for a provided type or collection of plants Applications of CYC • Pharmaceutical Term Thesaurus Manager/Integrator (Glaxo) • Terrorism Knowledge Base • MathCraft – 6th grade level help in understanding math Lots of Opinions on CYC/cyc/Cyc but true to GOFAI NELL • Never Ending Language Learner, Tom Mitchell, Carnegie Mellon University • We will never truly understand machine or human learning until we can build computer programs that, like people: – learn many different types of knowledge or functions, – from years of diverse, mostly self-supervised experience, – in a staged curricular fashion, where previously learned knowledge enables learning further types of knowledge, – where self-reflection and the ability to formulate new representations and new learning tasks enable the learner to avoid stagnation and performance plateaus. • T. Mitchell, et.al. AAAI 2015 NELL-2 • Given: – an initial ontology defining categories (e.g., Sport, Athlete) and Binary relations (e.g., AthletePlaysSport(x,y)), – approximately a dozen labeled training examples for each category and relation (e.g., examples of Sport might include the noun phrases “BaseBall” and “soccer”), – the weB (an initial 500 million weB pages from the ClueWeb 2009 collection (Callan and Hoy 2009), and access to 100,000 Google API search queries each day), – occasional interaction with humans (e.g., through NELL’s puBlic weBsite http://rtw.ml.cmu.edu); NELL -3 • DO: – Run 24 hours/day, forever, and each day: • 1. read (extract) more beliefs from the web, and remove old incorrect beliefs, to populate a growing Knowledge base containing a confidence and provenance for each belief, • 2. learn to read better than the previous day. NELL-4 Knowledge Base uses Blackboard Architecture Mitchell, et.al. (2015) NELL Results High confidence indicates that one of NELL’s modules assigns a 0.9 confidence to the belief or multiple modules propose the belief Mitchell, et.al. 2015 Alexa • Siri, Google Assistant too AWS Technologies Semantic Web 29 Semantic Web Big Data, API/Platform, & M2M Intersection •What is Semantic Web? •A collaborate movement led by W3C to promote common data format aimed at converting the current web dominated by unstructured and semi-structured document into a “web of data” •Components of Semantic Web • Resource Description Framework (RDF) • RDF Schema • Web Ontology Language (OWL) •Implications of Semantic Web • Link heterogeneous data sources on web • Drive “meaning” of information on the web • Enable machine-readable web • Enable intelligent services • Improve quality of search, targeted ads, etc. 30 Early Adopters of Semantic Web Synergy with Effortless Customer Experince schema.org project: a joint effort of Google, Yahoo!, and Microsoft documents the concepts and attributes that these three search engines will look for when Web publishers include Semantic Web data on their sites. That is RDF like data. Typical Applications Early Adopters • Optimized search engine • Knowledge graph • Knowledge management • Speech recognition • Content management • Customer/product driven analytics • Customer acquisition & retention • Predictive analytics for security • Visual discovery - 30% increase in traffic (Best Buy). 15% increase in CTR (Yahoo!). External Adopters - Semantic Technology In The Wild • Department of Defense • US Intelligence Agencies • Library of Congress • Wikipedia - DBpedia • NASA • Amdocs • BestBuy • BBC • MITRE Internet Semantic Web Linked Data Semantic Web – Background • Tim Berners-Lee 1999 • “I have a dream for the Web [in which computers] become capable of analyzing all the data on the Web – the content, links, and transactions between people and computers. A ‘Semantic Web’, which should make this possible, has yet to emerge, but when it does, the day-to-day mechanisms of trade, bureaucracy and our daily lives will be handled by machines talking to machines. The ‘intelligent agents’ people have touted for ages will finally materialize” • The dream is beginning to emerge and it’s all based on W3C Standards • Alias: Web of Data Web 3.0 • Linked Data Machine Readable Web • Our goal is to understand this concept, it’s formats and technologies and begin to explore some of the uses and power of this technology. SemTech Terms • Resource Description Framework (RDF) • RDF Schema (RDFS) • OWL (Web Ontology Language) • SPARQL • TripleStore - GraphDatabase • InFerring • Metadata • RDFa • MicroFormats • Ontology Modeling Today – Data and Information • Web Search: q Keyword Based “Results” q Provides a List of Links q Answer to Question? • Business Applications: q Query Based results q Range of information
Recommended publications
  • A Survey of Top-Level Ontologies to Inform the Ontological Choices for a Foundation Data Model
    A survey of Top-Level Ontologies To inform the ontological choices for a Foundation Data Model Version 1 Contents 1 Introduction and Purpose 3 F.13 FrameNet 92 2 Approach and contents 4 F.14 GFO – General Formal Ontology 94 2.1 Collect candidate top-level ontologies 4 F.15 gist 95 2.2 Develop assessment framework 4 F.16 HQDM – High Quality Data Models 97 2.3 Assessment of candidate top-level ontologies F.17 IDEAS – International Defence Enterprise against the framework 5 Architecture Specification 99 2.4 Terminological note 5 F.18 IEC 62541 100 3 Assessment framework – development basis 6 F.19 IEC 63088 100 3.1 General ontological requirements 6 F.20 ISO 12006-3 101 3.2 Overarching ontological architecture F.21 ISO 15926-2 102 framework 8 F.22 KKO: KBpedia Knowledge Ontology 103 4 Ontological commitment overview 11 F.23 KR Ontology – Knowledge Representation 4.1 General choices 11 Ontology 105 4.2 Formal structure – horizontal and vertical 14 F.24 MarineTLO: A Top-Level 4.3 Universal commitments 33 Ontology for the Marine Domain 106 5 Assessment Framework Results 37 F. 25 MIMOSA CCOM – (Common Conceptual 5.1 General choices 37 Object Model) 108 5.2 Formal structure: vertical aspects 38 F.26 OWL – Web Ontology Language 110 5.3 Formal structure: horizontal aspects 42 F.27 ProtOn – PROTo ONtology 111 5.4 Universal commitments 44 F.28 Schema.org 112 6 Summary 46 F.29 SENSUS 113 Appendix A F.30 SKOS 113 Pathway requirements for a Foundation Data F.31 SUMO 115 Model 48 F.32 TMRM/TMDM – Topic Map Reference/Data Appendix B Models 116 ISO IEC 21838-1:2019
    [Show full text]
  • Artificial Intelligence
    BROAD AI now and later Michael Witbrock, PhD University of Auckland Broad AI Lab @witbrock Aristotle (384–322 BCE) Organon ROOTS OF AI ROOTS OF AI Santiago Ramón y Cajal (1852 -1934) Cerebral Cortex WHAT’S AI • OLD definition: AI is everything we don’t yet know how program • Now some things that people can’t do: • unique capabilities (e.g. Style transfer) • superhuman performance (some areas of speech, vision, games, some QA, etc) • Current AI Systems can be divided by their kind of capability: • Skilled (Image recognition, Game Playing (Chess, Atari, Go, DoTA), Driving) • Attentive (Trading: Aidyia; Senior Care: CareMedia, Driving) • Knowledgeable, (Google Now, Siri, Watson, Cortana) • High IQ (Cyc, Soar, Wolfram Alpha) GOFAI • Thought is symbol manipulation • Large numbers of precisely defined symbols (terms) • Based on mathematical logic (implies (and (isa ?INST1 LegalAgreement) (agreeingAgents ?INST1 ?INST2)) (isa ?INST2 LegalAgent)) • Problems solved by searching for transformations of symbolic representations that lead to a solution Slow Development Thinking Quickly Thinking Slowly (System I) (System II) Human Superpower c.f. other Done well by animals and people animals Massively parallel algorithms Serial and slow Done poorly until now by computers Done poorly by most people Not impressive to ordinary people Impressive (prizes, high pay) "Sir, an animal’s reasoning is like a dog's walking on his hind legs. It is not done well; but you are surprised to find it done at all.“ - apologies to Samuel Johnson Achieved on computers by high- Fundamental design principle of power, low density, slow computers simulation of vastly different Computer superpower c.f. neural hardware human Recurrent Deep Learning & Deep Reasoning MACHINE LEARNING • Meaning is implicit in the data • Thought is the transformation of learned representations http://karpathy.github.io/2015/05/21/rnn- effectiveness/ .
    [Show full text]
  • Semantic Web Foundations for Representing, Reasoning, and Traversing Contextualized Knowledge Graphs
    Wright State University CORE Scholar Browse all Theses and Dissertations Theses and Dissertations 2017 Semantic Web Foundations for Representing, Reasoning, and Traversing Contextualized Knowledge Graphs Vinh Thi Kim Nguyen Wright State University Follow this and additional works at: https://corescholar.libraries.wright.edu/etd_all Part of the Computer Engineering Commons, and the Computer Sciences Commons Repository Citation Nguyen, Vinh Thi Kim, "Semantic Web Foundations for Representing, Reasoning, and Traversing Contextualized Knowledge Graphs" (2017). Browse all Theses and Dissertations. 1912. https://corescholar.libraries.wright.edu/etd_all/1912 This Dissertation is brought to you for free and open access by the Theses and Dissertations at CORE Scholar. It has been accepted for inclusion in Browse all Theses and Dissertations by an authorized administrator of CORE Scholar. For more information, please contact [email protected]. Semantic Web Foundations for Representing, Reasoning, and Traversing Contextualized Knowledge Graphs A dissertation submitted in partial fulfilment of the requirements for the degree of Doctor of Philosophy By VINH THI KIM NGUYEN B.E., Vietnam National University, 2007 2017 Wright State University WRIGHT STATE UNIVERSITY GRADUATE SCHOOL December 15, 2017 I HEREBY RECOMMEND THAT THE THESIS PREPARED UNDER MY SUPERVISION BY Vinh Thi Kim Nguyen ENTITLED Semantic Web Foundations of Representing, Reasoning, and Traversing in Contextualized Knowledge Graphs BE ACCEPTED IN PARTIAL FUL- FILLMENT OF THE REQUIREMENTS FOR THE DEGREE OF Doctor of Philosophy. Amit Sheth, Ph.D. Dissertation Director Michael Raymer, Ph.D. Director, Computer Science and Engineering Ph.D. Program Barry Milligan, Ph.D. Interim Dean of the Graduate School Committee on Final Examination Amit Sheth, Ph.D.
    [Show full text]
  • Knowledge Graphs on the Web – an Overview Arxiv:2003.00719V3 [Cs
    January 2020 Knowledge Graphs on the Web – an Overview Nicolas HEIST, Sven HERTLING, Daniel RINGLER, and Heiko PAULHEIM Data and Web Science Group, University of Mannheim, Germany Abstract. Knowledge Graphs are an emerging form of knowledge representation. While Google coined the term Knowledge Graph first and promoted it as a means to improve their search results, they are used in many applications today. In a knowl- edge graph, entities in the real world and/or a business domain (e.g., people, places, or events) are represented as nodes, which are connected by edges representing the relations between those entities. While companies such as Google, Microsoft, and Facebook have their own, non-public knowledge graphs, there is also a larger body of publicly available knowledge graphs, such as DBpedia or Wikidata. In this chap- ter, we provide an overview and comparison of those publicly available knowledge graphs, and give insights into their contents, size, coverage, and overlap. Keywords. Knowledge Graph, Linked Data, Semantic Web, Profiling 1. Introduction Knowledge Graphs are increasingly used as means to represent knowledge. Due to their versatile means of representation, they can be used to integrate different heterogeneous data sources, both within as well as across organizations. [8,9] Besides such domain-specific knowledge graphs which are typically developed for specific domains and/or use cases, there are also public, cross-domain knowledge graphs encoding common knowledge, such as DBpedia, Wikidata, or YAGO. [33] Such knowl- edge graphs may be used, e.g., for automatically enriching data with background knowl- arXiv:2003.00719v3 [cs.AI] 12 Mar 2020 edge to be used in knowledge-intensive downstream applications.
    [Show full text]
  • Using Linked Data for Semi-Automatic Guesstimation
    Using Linked Data for Semi-Automatic Guesstimation Jonathan A. Abourbih and Alan Bundy and Fiona McNeill∗ [email protected], [email protected], [email protected] University of Edinburgh, School of Informatics 10 Crichton Street, Edinburgh, EH8 9AB, United Kingdom Abstract and Semantic Web systems. Next, we outline the process of GORT is a system that combines Linked Data from across guesstimation. Then, we describe the organisation and im- several Semantic Web data sources to solve guesstimation plementation of GORT. Finally, we close with an evaluation problems, with user assistance. The system uses customised of the system’s performance and adaptability, and compare inference rules over the relationships in the OpenCyc ontol- it to several other related systems. We also conclude with a ogy, combined with data from DBPedia, to reason and per- brief section on future work. form its calculations. The system is extensible with new Linked Data, as it becomes available, and is capable of an- Literature Survey swering a small range of guesstimation questions. Combining facts to answer a user query is a mature field. The DEDUCOM system (Slagle 1965) was one of the ear- Introduction liest systems to perform deductive query answering. DE- The true power of the Semantic Web will come from com- DUCOM applies procedural knowledge to a set of facts in bining information from heterogeneous data sources to form a knowledge base to answer user queries, and a user can new knowledge. A system that is capable of deducing an an- also supplement the knowledge base with further facts.
    [Show full text]
  • Why Has AI Failed? and How Can It Succeed?
    Why Has AI Failed? And How Can It Succeed? John F. Sowa VivoMind Research, LLC 10 May 2015 Extended version of slides for MICAI'14 ProblemsProblems andand ChallengesChallenges Early hopes for artificial intelligence have not been realized. Language understanding is more difficult than anyone thought. A three-year-old child is better able to learn, understand, and generate language than any current computer system. Tasks that are easy for many animals are impossible for the latest and greatest robots. Questions: ● Have we been using the right theories, tools, and techniques? ● Why haven’t these tools worked as well as we had hoped? ● What other methods might be more promising? ● What can research in neuroscience and psycholinguistics tell us? ● Can it suggest better ways of designing intelligent systems? 2 Early Days of Artificial Intelligence 1960: Hao Wang’s theorem prover took 7 minutes to prove all 378 FOL theorems of Principia Mathematica on an IBM 704 – much faster than two brilliant logicians, Whitehead and Russell. 1960: Emile Delavenay, in a book on machine translation: “While a great deal remains to be done, it can be stated without hesitation that the essential has already been accomplished.” 1965: Irving John Good, in speculations on the future of AI: “It is more probable than not that, within the twentieth century, an ultraintelligent machine will be built and that it will be the last invention that man need make.” 1968: Marvin Minsky, technical adviser for the movie 2001: “The HAL 9000 is a conservative estimate of the level of artificial intelligence in 2001.” 3 The Ultimate Understanding Engine Sentences uttered by a child named Laura before the age of 3.
    [Show full text]
  • Chaudron: Extending Dbpedia with Measurement Julien Subercaze
    Chaudron: Extending DBpedia with measurement Julien Subercaze To cite this version: Julien Subercaze. Chaudron: Extending DBpedia with measurement. 14th European Semantic Web Conference, Eva Blomqvist, Diana Maynard, Aldo Gangemi, May 2017, Portoroz, Slovenia. hal- 01477214 HAL Id: hal-01477214 https://hal.archives-ouvertes.fr/hal-01477214 Submitted on 27 Feb 2017 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Chaudron: Extending DBpedia with measurement Julien Subercaze1 Univ Lyon, UJM-Saint-Etienne, CNRS Laboratoire Hubert Curien UMR 5516, F-42023, SAINT-ETIENNE, France [email protected] Abstract. Wikipedia is the largest collaborative encyclopedia and is used as the source for DBpedia, a central dataset of the LOD cloud. Wikipedia contains numerous numerical measures on the entities it describes, as per the general character of the data it encompasses. The DBpedia In- formation Extraction Framework transforms semi-structured data from Wikipedia into structured RDF. However this extraction framework of- fers a limited support to handle measurement in Wikipedia. In this paper, we describe the automated process that enables the creation of the Chaudron dataset. We propose an alternative extraction to the tra- ditional mapping creation from Wikipedia dump, by also using the ren- dered HTML to avoid the template transclusion issue.
    [Show full text]
  • Logic-Based Technologies for Intelligent Systems: State of the Art and Perspectives
    information Article Logic-Based Technologies for Intelligent Systems: State of the Art and Perspectives Roberta Calegari 1,* , Giovanni Ciatto 2 , Enrico Denti 3 and Andrea Omicini 2 1 Alma AI—Alma Mater Research Institute for Human-Centered Artificial Intelligence, Alma Mater Studiorum–Università di Bologna, 40121 Bologna, Italy 2 Dipartimento di Informatica–Scienza e Ingegneria (DISI), Alma Mater Studiorum–Università di Bologna, 47522 Cesena, Italy; [email protected] (G.C.); [email protected] (A.O.) 3 Dipartimento di Informatica–Scienza e Ingegneria (DISI), Alma Mater Studiorum–Università di Bologna, 40136 Bologna, Italy; [email protected] * Correspondence: [email protected] Received: 25 February 2020; Accepted: 18 March 2020; Published: 22 March 2020 Abstract: Together with the disruptive development of modern sub-symbolic approaches to artificial intelligence (AI), symbolic approaches to classical AI are re-gaining momentum, as more and more researchers exploit their potential to make AI more comprehensible, explainable, and therefore trustworthy. Since logic-based approaches lay at the core of symbolic AI, summarizing their state of the art is of paramount importance now more than ever, in order to identify trends, benefits, key features, gaps, and limitations of the techniques proposed so far, as well as to identify promising research perspectives. Along this line, this paper provides an overview of logic-based approaches and technologies by sketching their evolution and pointing out their main application areas. Future perspectives for exploitation of logic-based technologies are discussed as well, in order to identify those research fields that deserve more attention, considering the areas that already exploit logic-based approaches as well as those that are more likely to adopt logic-based approaches in the future.
    [Show full text]
  • Application of Ontological Models for Representing Engineering Concepts in Engineering
    SCIENTIFIC PROCEEDINGS X INTERNATIONAL CONGRESS "MACHINES, TECHNOLОGIES, MATERIALS" 2013 ISSN 1310-3946 APPLICATION OF ONTOLOGICAL MODELS FOR REPRESENTING ENGINEERING CONCEPTS IN ENGINEERING Martin Ivanov1, Nikolay Tontchev2 New Bulgarian University, Sofia, Montivideo 211, University of Transport, Sofia, Geo Milev 1582 Abstract: The necessity, opportunities and advantages from developing ontological models of engineering concepts in the field of machine manufacturing technology are introduced. The characteristics and principles of the methodology for developing engineering technology models are presented. An experimental ontological model developed by the authors based on Protégé OWL 4.2 is described. Keywords: ONTOLOGICAL MODELS, ENGINEERING CONCEPTS, IDEF5, PROTEGE OWL 1. Introduction - Existing objects and phenomena in the subject area in the terms of the accepted terminology (vocabulary/vocabularies, Modern engineering activities take place in an evolving and in this case – the so called engineering lexicon or EL); enriching information environment characterized by exponentially - How objects and phenomena relate to each other; increasing information flows, variety of means for access and in the formats of data presentation, diversity of engineering problems, - How they are used within the subject area, and outside it; solved by automated means, penetration of the artificial intelligence methods and the semantic Web, too. Effectiveness of engineering is - Rules that define their existence and behavior. highly dependent on the access and the opportunities for efficient The ontology is composed on the basis of a specific vocabulary processing of this information, in particular by the use of models for and terminology characteristic for the subject area. This implies (a) formally describing, processing and dissemination of knowledge strict definitions (axioms) of the terms and rules (grammars) resulting from research and engineering practice.
    [Show full text]
  • Knowledge on the Web: Towards Robust and Scalable Harvesting of Entity-Relationship Facts
    Knowledge on the Web: Towards Robust and Scalable Harvesting of Entity-Relationship Facts Gerhard Weikum Max Planck Institute for Informatics http://www.mpi-inf.mpg.de/~weikum/ Acknowledgements 2/38 Vision: Turn Web into Knowledge Base comprehensive DB knowledge fact of human knowledge assets extraction • everything that (Semantic (Statistical Web) Web) Wikipedia knows • machine-readable communities • capturing entities, (Social Web) classes, relationships Source: DB & IR methods for knowledge discovery. Communications of the ACM 52(4), 2009 3/38 Knowledge as Enabling Technology • entity recognition & disambiguation • understanding natural language & speech • knowledge services & reasoning for semantic apps • semantic search: precise answers to advanced queries (by scientists, students, journalists, analysts, etc.) German chancellor when Angela Merkel was born? Japanese computer science institutes? Politicians who are also scientists? Enzymes that inhibit HIV? Influenza drugs for pregnant women? ... 4/38 Knowledge Search on the Web (1) Query: sushi ingredients? Results: Nori seaweed Ginger Tuna Sashimi ... Unagi http://www.google.com/squared/5/38 Knowledge Search on the Web (1) Query:Query: JapaneseJapanese computerscomputeroOputer science science ? institutes ? http://www.google.com/squared/6/38 Knowledge Search on the Web (2) Query: politicians who are also scientists ? ?x isa politician . ?x isa scientist Results: Benjamin Franklin Zbigniew Brzezinski Alan Greenspan Angela Merkel … http://www.mpi-inf.mpg.de/yago-naga/7/38 Knowledge Search on the Web (2) Query: politicians who are married to scientists ? ?x isa politician . ?x isMarriedTo ?y . ?y isa scientist Results (3): [ Adrienne Clarkson, Stephen Clarkson ], [ Raúl Castro, Vilma Espín ], [ Jeannemarie Devolites Davis, Thomas M. Davis ] http://www.mpi-inf.mpg.de/yago-naga/8/38 Knowledge Search on the Web (3) http://www-tsujii.is.s.u-tokyo.ac.jp/medie/ 9/38 Take-Home Message If music was invented Information is not Knowledge.
    [Show full text]
  • Talking to Computers in Natural Language
    feature Talking to Computers in Natural Language Natural language understanding is as old as computing itself, but recent advances in machine learning and the rising demand of natural-language interfaces make it a promising time to once again tackle the long-standing challenge. By Percy Liang DOI: 10.1145/2659831 s you read this sentence, the words on the page are somehow absorbed into your brain and transformed into concepts, which then enter into a rich network of previously-acquired concepts. This process of language understanding has so far been the sole privilege of humans. But the universality of computation, Aas formalized by Alan Turing in the early 1930s—which states that any computation could be done on a Turing machine—offers a tantalizing possibility that a computer could understand language as well. Later, Turing went on in his seminal 1950 article, “Computing Machinery and Intelligence,” to propose the now-famous Turing test—a bold and speculative method to evaluate cial intelligence at the time. Daniel Bo- Figure 1). SHRDLU could both answer whether a computer actually under- brow built a system for his Ph.D. thesis questions and execute actions, for ex- stands language (or more broadly, is at MIT to solve algebra word problems ample: “Find a block that is taller than “intelligent”). While this test has led to found in high-school algebra books, the one you are holding and put it into the development of amusing chatbots for example: “If the number of custom- the box.” In this case, SHRDLU would that attempt to fool human judges by ers Tom gets is twice the square of 20% first have to understand that the blue engaging in light-hearted banter, the of the number of advertisements he runs, block is the referent and then perform grand challenge of developing serious and the number of advertisements is 45, the action by moving the small green programs that can truly understand then what is the numbers of customers block out of the way and then lifting the language in useful ways remains wide Tom gets?” [1].
    [Show full text]
  • Conceptualization and Visual Knowledge Organization: a Survey of Ontology-Based Solutions
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Repository of the Academy's Library CONCEPTUALIZATION AND VISUAL KNOWLEDGE ORGANIZATION: A SURVEY OF ONTOLOGY-BASED SOLUTIONS A. Benedek1, G. Lajos2 1 Hungarian Academy of Sciences (HUNGARY) 2 WikiNizer (UNITED KINGDOM) Abstract Conceptualization and Visual Knowledge Organization are overlapping research areas of Intellect Augmentation with hot topics at their intersection which are enhanced by proliferating issues of ontology integration. Knowledge organization is more closely related to the content of concepts and to the actual knowledge items than taxonomic structures and ‘ontologies’ understood as “explicit specifications of a conceptualization” (Gruber). Yet, the alignment of the structure and relationship of knowledge items of a domain of interest require similar operations. We reconsider ontology-based approaches to collaborative conceptualization from the point of view of user interaction in the context of augmented visual knowledge organization. After considering several formal ontology based approaches, it is argued that co-evolving visualization of concepts are needed both at the object and meta-level to handle the social aspects of learning and knowledge work in the framework of an Exploratory Epistemology. Current systems tend to separate the conceptual level from the content, and the context of logical understanding from the intent and use of concepts. Besides, they usually consider the unification of individual contributions in a ubiquitous global representation. As opposed to this God's eye view we are exploring new alternatives that aim to provide Morphic-like live, tinkerable, in situ approaches to conceptualization and its visualization and to direct manipulation based ontology authoring capabilities.
    [Show full text]