Foundations of Ontology Reconciliation

Total Page:16

File Type:pdf, Size:1020Kb

Foundations of Ontology Reconciliation Chapter 3 Foundations of Ontology Reconciliation Concepts without percepts are empty, percepts without concepts are blind. Immanuel Kant (A critique of pure reason, 1781) This chapter aims at providing a theoretical basis for ontology reconciliation. This leads to philosophical topics such as what meaning is and how it can be represented in an ontology. As innocent as these issues might appear, they open a Pandora’s Box of problems. Our discussion is limited to those issues that are relevant for the current investigation. Section 3.1 presents a brief analysis of meaning from a semiotic perspective. Section 3.2 discusses how meaning can be satisfactorily represented. In particular, we describe Tarskian semantics, its shortcomings, and a way to overcome these problems. Section 3.3 elaborates on these ideas and argues how automatic ontology mapping can be established. Furthermore, we compare our approach to ontology reconciliation with other approaches. We conclude in Section 3.4. 3.1 Meaning We start our discussion with an analysis of meaning in the context of semiotics, i.e. the study of signs, symbols and meanings. One of the founders of semiotics is Charles Sanders Peirce (1839-1914). His account of meaning is discussed below. 3.1.1 The Peircean Model In the model of Peirce, a sign is assumed to consist of three parts, yielding a triadic model (Chandler, 2002). Whereas many authors have accepted the presence of three 40 Foundations of Ontology Reconciliation thought or reference symbol referent Figure 3.1: The Peircean model of a sign components of a sign, there is no consensus on the names of these components and their exact details. In our discussion, we adopt the names that are used by Ogden and Richards (1923) and try to stay away from as many controversial issues as possible. Peirce’s triadic model is often represented as a triangle with the three compo- nents at its corners, as depicted in Figure 3.1. The first part of the sign is the symbol and regards the form which the sign takes. The second part is the thought or reference, which regards the sense made of a sign. The third part is the referent, i.e. that to which the sign refers. Each of the three components is discussed in more depth below. Symbol A symbol can be many things, such as a word, an image or a sound. When someone interprets a symbol as standing for something else, the symbol can be viewed as one of the three components that make up a sign. Many semioticians, among whom de Saussure and Peirce, have argued that symbols may be completely arbitrary, i.e. no natural connection may exist between a symbol and what it stands for. For this reason, Figure 3.1 shows the link between the symbol and the referent as a dashed line. According to Peirce, signs may be arbitrary in different degrees. An example of a sign that is not completely arbitrary is an icon. An icon is a sign in which the symbol resembles what it stands for in some way. An example of an icon is , commonly used for representing an airport. This sign qualifies as an icon for airports because it resembles, in some sense, the shape of an airplane. Other examples of icons are onomatopoeia, i.e. words that sound similar to what they refer to, such as "cuckoo" and "boom". Nevertheless, most symbols in a language are not iconic, but, as Peirce calls it, symbolic. A word such as "airplane" bears no natural relation to a real airplane, not in shape, neither in sound. Also, the concept-names in an ontology and the bit strings in a computer should be regarded as completely symbolic, i.e. as completely arbitrary signs. 3.1 Meaning 41 Thought or reference If the relation between the symbol and its referent is not established by resemblance, how else is it established? According to Peirce, the symbol and referent are related because they are mediated by the thought or reference (or interpretant in his termi- nology). The thought or reference corresponds to the sense that is made of the sign. Therefore, the theory presumes the presence of a sense-maker, or an interpreter. Thus, the symbol is connected to its referent by the thought of its interpreter. Meaning arises in the Peircean model, when the interpreter makes sense of the symbol. This process is called semiosis. The meaning that arises during semiosis can be viewed as a relation between the symbol and the referent, i.e. corresponding to thought or reference. Therefore, other versions of the semiotic triangle have used the word "meaning" to refer to what we call "thought or reference"(Vogt, 2002). Referent The referent is the object, in some kind of external and objective reality, to which the symbol refers. Objects are not restricted to concrete objects, but may be interpreted rather broadly. Referents may be abstract, fictional or refer to a state of affairs. For example, a traffic light sign for stop consists of: a red light facing traffic at an intersection (the symbol), the idea that a red light indicates that vehicles must stop (the thought or reference) and vehicles halting (the referent) (Chandler, 2002). 3.1.2 Sinn und Bedeutung Gottlob Frege (1848-1925) was a German philosopher who has been very influential in the philosophy of language. In his seminal paper "Über Sinn und Bedeutung" (Frege, 1892), Frege argued that the referent of a term alone is insufficient to provide a complete account of meaning. His argument is as follows. Consider the terms morning star and evening star which both have the same referent (or bedeutung in his terminology), namely Venus. If the meaning of a term were fully determined by its referent, then the morning star would mean exactly the same as the evening star.A consequence of this would be that the sentence "the morning star is the evening star" would mean exactly the same as the sentence "the morning star is the morning star". However, the first sentence describes a major astronomical discovery, whereas the last sentence is trivial, i.e. it is true on the basis of its constituents. Frege proposed to solve this problem by adding an extra component to the analysis of meaning, namely sinn. Whereas, the morning star and the evening star have the same referent (bedeutung), they differ in sinn. Consequently, the sentence "the morning star is the evening star" can be explained to differ in meaning from the sentence "the morning star is the morning star". Frege’s distinction between sinn and bedeutung matches well with the Peircean triad (Sowa, 2000, p.192). Frege’s sinn corresponds to Peirce’s thought or reference and Frege’s bedeutung corresponds to Peirce’s referent. In Peirce’s model, the terms morning star and evening star have the same referent. The thought or reference differs 42 Foundations of Ontology Reconciliation as the first term evokes a thought of being visible in the morning and the latter term evokes a thought of being visible in the evening. 3.1.3 Conventionality The semiotic analysis described in this section has stressed once more the need for conventions in symbolic systems. The more arbitrary a sign is, the more important the role of conventions. Whereas iconic signs involve some degree of conven- tionality, communication using symbolic signs is based on convention. This social dimension has been recognized by most semioticians. Whereas humans usually interpret signs by unconsciously relating them to a familiar system of conventions (Chandler, 2002), computer systems must be pro- vided with an explicit account of conventions. Ontologies have been proposed to make such conventions explicit. Thus, the purpose of ontologies in IS can be given a justification from semiotics. The next section discusses how ontologies can be realized in a way that is semiotically justified. In particular, we will discuss how meaning can be represented in a computer. 3.2 Meaning Representation The following quotation characterizes the problem that is currently encountered by AI researchers doing knowledge representation. The notation we use must be understandable to those using it and read- ing it; so it must have a semantics; so it must have a Tarskian semantics, because there is no other candidate. (McDermott, 1987) As the above quotation indicates, Tarskian semantics is the most common way to provide a formal account of meaning. This also appears from formalisms that have been proposed to develop ontologies, which mostly follow a Tarskian approach. Tarskian semantics is discussed in Section 3.2.1 and 3.2.2. Despite its popularity in the KR community, Tarskian semantics is not ideal for all purposes. In fact, it is difficult to justify from a semiotic perspective. We will discuss two problems of Tarskian semantics, namely the problem of indistinguish- able models in Section 3.2.3 and the symbol grounding problem in Section 3.2.4. A way to overcome these problems is discussed in Section 3.2.5. 3.2.1 World As described in a well-known textbook on AI (Genesereth and Nilsson, 1987), the formalization of knowledge begins with determining which objects are assumed to exist and their interrelationships. The set of objects in the world that is inhabited by the agent is called the domain of discourse, denoted by ∆I. There is usually not 3.2 Meaning Representation 43 one unique way to determine which objects exist in the world. Basically, an object may be anything about which we want to say something. For an agent to be able to store knowledge about ∆I, it needs a conceptualiza- tion, denoted by ρ.
Recommended publications
  • THE SYMBOL GROUNDING PROBLEM Stevan HARNAD
    Physica D 42 (1990) 335-346 North-Holland THE SYMBOL GROUNDING PROBLEM Stevan HARNAD Department of Psychology, Princeton University, Princeton, NJ 08544, USA There has been much discussion recently about the scope and limits of purely symbolic models of the mind and abotlt the proper role of connectionism in cognitive modeling. This paper describes the "symbol grounding problem": How can the semantic interpretation of a formal symbol system be made intrinsic to the system, rather than just parasitic on the meanings in our heads? How can the meanings of the meaningless symbol tokens, manipulated solely on the basis of their (arbitrary) shapes, be grounded in anything but other meaningless symbols? The problem is analogous to trying to learn Chinese from a Chinese/Chinese dictionary alone. A candidate solution is sketched: Symbolic representations must be grounded bottom-up in nonsymbolic representations of two kinds: (1) iconic representations, which are analogs of the proximal sensory projections of distal objects and events, and (2) categorical representations, which are learned and innate feature detectors that pick out the invariant features of object and event categories from their sensory projections. Elementary symbols are the names of these object and event categories, assigned on the basis of their (nonsymbolic) categorical representations. Higher-order (3) symbolic representations, grounded in these elementary symbols, consist of symbol strings describing category membership relations (e.g. "An X is a Y that is Z "). Connectionism is one natural candidate for the mechanism that learns the invariant features underlying categorical representations, thereby connecting names to the proximal projections of the distal objects they stand for.
    [Show full text]
  • The Symbol Grounding Problem ... Remains Unsolved∗
    The Symbol Grounding Problem ::: Remains Unsolved∗ Selmer Bringsjord Department of Cognitive Science Department of Computer Science Lally School of Management & Technology Rensselaer AI & Reasoning (RAIR) Laboratory Rensselaer Polytechnic Institute (RPI) Troy NY 12180 USA [email protected] version of 082613NY Contents 1 Introduction and Plan 1 2 What is the Symbol Grounding Problem? 1 3 T&F's Proposed Solution 5 3.1 Remembering Cog . .7 4 The Refutation 8 4.1 Anticipated Rebuttals . 10 Rebuttal 1 . 11 Rebuttal 2 . 12 5 Concluding Remarks 12 References 15 List of Figures 1 A Concretization of the SGP . .5 2 Floridi's Ontology of Information . .9 ∗I'm indebted to Stevan Harnad for sustained and unforgettable discussions both of SGP and Searle's Chinese Room Argument (CRA), and to Luciano Floridi for teaching me enough about information to realize that SGP centers around the level of semantic information at the heart of CRA. 1 Introduction and Plan Taddeo & Floridi (2007) propose a solution to the Symbol Grounding Prob- lem (SGP).1 Unfortunately, their proposal, while certainly innovative and interesting, merely shows that a class of robots can in theory connect, in some sense, the symbols it manipulates with the external world it perceives, and can, on the strength of that connection, communicate in sub-human fashion.2 Such a connection-communication combination is routinely forged (by e.g. the robots in my laboratory (by the robot PERI in Bringsjord & Schimanski 2003) and in countless others, and indeed by robots now on the market), but is rendered tiny and insignificant in the vast and towering face of SGP, which is the problem of building a robot that genuinely understands semantic information at the level of human (natural) language, and that ac- quires for itself the knowledge we gain from reading such language.
    [Show full text]
  • Modelling Cognition
    Modelling Cognition SE 367 : Cognitive Science Group C Nature of Linguistic Sign • Linguistic sign – Not - Thing to Name – Signified and Signifier – The semantic breaking is arbitrary • Ex. The concept of eat and drink in Bengali being mapped to the same sound-image NATURE OF THE LINGUISTIC SIGN -- FERDINAND DE SAUSSURE The Sign • Icon – In the mind – Existence of the ‘object’ – not necessary • Index – Dynamic connection to the object by blind compulsion – If the object ceases to exist, the index loses its significance • Symbol – Medium of communication THE SIGN: ICON, INDEX AND SYMBOL -- CHARLES SANDERS PEIRCE Symbol Grounding Problem and Symbolic Theft • Chinese-Chinese dictionary recursion • Symbolic representations (to be grounded) • Non symbolic representations (sensory) – Iconic – Categorical THE SYMBOL GROUNDING PROBLEM -- Stevan Harnad Symbol Grounding Problem and Symbolic Theft • Symbol Systems – Higher level cognition – semantics • Connectionist systems – Capture invariant features – Identification and discrimination • Sensorimeter toil • Symbolic theft Symbol Grounding and Symbolic Theft Hypothesis– A.Cangelosi , A. Greco and S. Harnad A Computer Program ? • Computer Program –Searle – Chinese Room – Missing semantics – Compatibility of programs with any hardware – contrary to the human mind – Simulation vs. Duplication Is the Brain's Mind a Computer Program? -- John R. Searle Image Schema • A condensed description of perceptual experience for the purpose of mapping spatial structure onto conceptual structure. An Image Schema Language – RS Amant, CT Morrison, Yu-Han Chang, PR Cohen,CBeal Locating schemas in a cognitive architecture An Image Schema Language – RS Amant, CT Morrison, Yu-Han Chang, PR Cohen,CBeal Image Schema Language • Implementing image schemas gives us insight about how they can function as a semantic core for reasoning.
    [Show full text]
  • The Symbol Grounding Problem
    The Symbol Grounding Problem http://www.ecs.soton.ac.uk/~harnad/Papers/Harnad/harnad90.sgproble... Harnad, S. (1990) The Symbol Grounding Problem. Physica D 42: 335-346. THE SYMBOL GROUNDING PROBLEM Stevan Harnad Department of Psychology Princeton University Princeton NJ 08544 [email protected] ABSTRACT: There has been much discussion recently about the scope and limits of purely symbolic models of the mind and about the proper role of connectionism in cognitive modeling. This paper describes the "symbol grounding problem": How can the semantic interpretation of a formal symbol system be made intrinsic to the system, rather than just parasitic on the meanings in our heads? How can the meanings of the meaningless symbol tokens, manipulated solely on the basis of their (arbitrary) shapes, be grounded in anything but other meaningless symbols? The problem is analogous to trying to learn Chinese from a Chinese/Chinese dictionary alone. A candidate solution is sketched: Symbolic representations must be grounded bottom-up in nonsymbolic representations of two kinds: (1) "iconic representations" , which are analogs of the proximal sensory projections of distal objects and events, and (2) "categorical representations" , which are learned and innate feature-detectors that pick out the invariant features of object and event categories from their sensory projections. Elementary symbols are the names of these object and event categories, assigned on the basis of their (nonsymbolic) categorical representations. Higher-order (3) "symbolic representations" , grounded in these elementary symbols, consist of symbol strings describing category membership relations (e.g., "An X is a Y that is Z"). Connectionism is one natural candidate for the mechanism that learns the invariant features underlying categorical representations, thereby connecting names to the proximal projections of the distal objects they stand for.
    [Show full text]
  • Machine Symbol Grounding and Optimization
    MACHINE SYMBOL GROUNDING AND OPTIMIZATION Oliver Kramer International Computer Science Institute Berkeley [email protected] Keywords: autonomous agents, symbol grounding, zero semantical commitment condition, machine learning, interface design, optimization Abstract: Autonomous systems gather high-dimensional sensorimotor data with their multimodal sensors. Symbol grounding is about whether these systems can, based on this data, construct symbols that serve as a vehi- cle for higher symbol-oriented cognitive processes. Machine learning and data mining techniques are geared towards finding structures and input-output relations in this data by providing appropriate interface algorithms that translate raw data into symbols. Can autonomous systems learn how to ground symbols in an unsuper- vised way, only with a feedback on the level of higher objectives? A target-oriented optimization procedure is suggested as a solution to the symbol grounding problem. It is demonstrated that the machine learning perspective introduced in this paper is consistent with the philosophical perspective of constructivism. Inter- face optimization offers a generic way to ground symbols in machine learning. The optimization perspective is argued to be consistent with von Glasersfeld’s view of the world as a black box. A case study illustrates technical details of the machine symbol grounding approach. 1 INTRODUCTION However, how do we define symbols and their meaning in artificial systems, e.g., for autonomous robots? Which subsymbolic elements belong to the The literature on artificial intelligence (AI) defines set that defines a symbol, and – with regard to cog- “perception” in cognitive systems as the transduction nitive manipulations – what is the interpretation of of subsymbolic data to symbols (e.g.
    [Show full text]
  • Symbol Grounding in Computational Systems: a Paradox of Intentions’, Minds and Machines, 19 (4), 529-41
    Müller, Vincent C. (2009), ‘Symbol grounding in computational systems: A paradox of intentions’, Minds and Machines, 19 (4), 529-41. - The published version is earlier (2006) than this one. - http://doi.org/10.1007/s11023-009-9175-1 - http://www.sophia.de - http://orcid.org/0000-0002-4144-4957 Symbol Grounding in Computational Systems: A Paradox of Intentions Vincent C. Müller Anatolia College/ACT, P.O. Box 21021, 55510 Pylaia, Greece [email protected], http://www.typos.de Abstract. The paper presents a paradoxical feature of computational systems that suggests that computationalism cannot explain symbol grounding. If the mind is a digital computer, as computationalism claims, then it can be comput- ing either over meaningful symbols or over meaningless symbols. If it is com- puting over meaningful symbols its functioning presupposes the existence of meaningful symbols in the system, i.e. it implies semantic nativism. If the mind is computing over meaningless symbols, no intentional cognitive processes are available prior to symbol grounding; therefore no symbol grounding could take place since any such process presupposes intentional processes. So, whether computing in the mind is over meaningless or over meaningful symbols, com- putationalism implies semantic nativism. Keywords. Artificial intelligence, computationalism, language of thought, Fodor, Putnam, semantic nativism, symbol grounding, syntactic computation. 1. Computationalism We will discuss an aspect of the problems a digital computational system has to ac- quire meaningful symbols and what these problems mean for a computational theory of the mind, in particular its relation to semantic nativism. The classical computa- tional theory of the mind, or “computationalism” for short, holds that the mind is a digital computer, in particular that it is a computational information processor.
    [Show full text]
  • Semantics and Symbol Grounding in Turing Machine Processes
    CORE Metadata, citation and similar papers at core.ac.uk Provided by Analecta Cracoviensia Semina Nr 16 Scientiarum 2017 s. 211–223 DOI: http://dx.doi.org/10.15633/ss.2492 Anna Sarosiek Semantics and symbol grounding in Turing machine processes The symbol grounding problem is one of the most important ques- tions encountered in Artificial Intelligence research. The problem is very old and repeatedly discussed.1 It received special attention from Steven Harnad.2 There are many articles attempting to de- scribe the way to solve this problem.3 Despite this, works on Arti- ficial Intelligence still face obstacles to proper embedding of sym- bols of artificial systems. The difficulty of the symbol grounding problem is that we cannot accurately determine how to depict the content of a representative element in a symbolic system – how sensory data affects the whole of symbolic representations. Likewise, we do not know how sensory 1 See A. Cangelosi, A. Greco, S. Harnad, Symbol Grounding and the Symbolic Theft Hypothesis, in: Simulating the Evolution of Language, ed. A. Cangelosi, D. Pa- risi, London 2002; S. Harnad, The Symbol Grounding Problem, “Physica D: Nonli- near Phenomena” 42 (1990) iss. 1, p. 335–346; K. MacDorman, Cognitive Robotics. Grounding Symbols through Sensorimotor Integration, “Journal of the Robotics So- ciety of Japan” 17 (1999) no. 1, p. 20–24. 2 See Professor Stevan Harnad, http://www.ecs.soton.ac.uk/people/harnad (22.11.2017). 3 See e.g. L. Steels, The Symbol Grounding Problem Has Been Solved, so Wha- t’s Next, “Symbols and Embodiment: Debates on Meaning and Cognition” (2008), p.
    [Show full text]
  • Symbol Grounding Problem Angelo Loula1,2 & João Queiroz 2,3* 1 Informatics Area, Exact Sciences Department, State University of Feira De Santana, Brazil
    1 Symbol Grounding Problem Angelo Loula1,2 & João Queiroz 2,3* 1 Informatics Area, Exact Sciences Department, State University of Feira de Santana, Brazil. 2 Department of Computer Engineering and Industrial Automation, School of Electrical and Computer Engineering, State University of Campinas (UNICAMP), Brazil. 3 Research Group on History, Philosophy, and Biological Sciences Teaching, Institute of Biology, Federal University of Bahia, Brazil. * Corresponding Author [email protected] 1 THIS PREPRINT WAS PUBLISHED AS: Loula, A., Queiroz, J. (2008) Symbol grounding problem. In: Rabunal, J., Dorado, J., Pazos, A. (eds.) Encyclopedia of Artificial Intelligence, pp. 1543–1548. IGI Global INTRODUCTION The topic of representation acquisition, manipulation and use has been a major trend in Artificial Intelligence since its beginning and persists as an important matter in current research. Particularly, due to initial focus on development of symbolic systems, this topic is usually related to research in symbol grounding by artificial intelligent systems. Symbolic systems, as proposed by Newell & Simon (1976), are characterized as a high-level cognition system in which symbols are seen as “[lying] at the root of intelligent action” (Newell and Simon, 1976, p.83). Moreover, they stated the Physical Symbol Systems Hypothesis (PSSH), making the strong claim that “a physical symbol system has the necessary and sufficient means for general intelligent action” (p.87). This hypothesis, therefore, sets equivalence between symbol systems and intelligent action, in such a way that every intelligent action would be originated in a symbol system and every symbol system is capable of intelligent action. The symbol system described by Newell and Simon (1976) is seen as a computer program capable of manipulating entities called symbols, ‘physical patterns’ combined in expressions, which can be created, modified or destroyed by syntactic processes.
    [Show full text]
  • Towards a Model for Grounding Semantic Composition
    http://www.diva-portal.org This is the published version of a paper presented at AISB-Artificial Intelligence and the Simulation of Behaviour 2014. Citation for the original published paper: Daoutis, M., Mavridis, N. (2014) Towards a Model for Grounding Semantic Composition. In: Goldsmiths, University of London, United Kingdom N.B. When citing this work, cite the original published paper. Permanent link to this version: http://urn.kb.se/resolve?urn=urn:nbn:se:oru:diva-34663 Towards a Model for Grounding Semantic Composition Marios Daoutis1 and Nikolaos Mavridis2 Abstract. Compositionality is widely accepted as a fundamental prin- ciple in linguistics and is also acknowledged as a key cog- Concept nitive capacity. However, despite the prime importance of (Conceptualisation) compositionality towards explaining the nature of meaning and concepts in cognition, and despite the need for compu- represented by tational models which are able to process the composition of grounded meaning there is little existing research. Thus, we aim to create computational models that concern the semantic composition of grounded meaning, that can be applied to em- abstracted to bodied intelligent agents (such as cognitive robots), in order to make them capable of creating and processing grounded Ullmann perceptual-semantic associations, and most importantly their 1973 compositions, taking into account syntactic, pragmatic as well Thing Symbol refered by as semantic considerations. Here we focus on an introduction (Reality) (Language) to the problem, while then we review related work across mul- tiple directions. Finally we propose a set of concrete desiderata that a computational theory of grounded semantic composition for embodied agents should satisfy, thus paving a clear avenue Figure 1.
    [Show full text]
  • Syntactic Semantics and the Proper Treatment of Computationalism
    1 Syntactic Semantics and the Proper Treatment of Computationalism William J. Rapaport Department of Computer Science and Engineering, Department of Philosophy, Department of Linguistics, and Center for Cognitive Science University at Buffalo, The State University of New York Buffalo, NY 14260-2500 [email protected] http://www.cse.buffalo.edu/~rapaport Keywords: computationalism, semiotic systems, cognition, syntax, semantics 2 Abstract Computationalism should not be the view that (human) cognition is computation; it should be the view that cognition (simpliciter) is computable. It follows that computationalism can be true even if (human) cognition is not the result of computations in the brain. If semiotic systems are systems that interpret signs, then both humans and computers are semiotic systems. Finally, minds can be considered as virtual machines implemented in certain semiotic systems, primarily the brain, but also AI computers. 1. Introduction This essay treats three topics: computationalism, semiotic systems, and cognition (the mind), offering what I feel is the proper treatment of computationalism. From this, certain views about semiotic systems and minds follow (or, at least, are consistent): First, I argue that computationalism should not be understood as the view that (human) cognition is computation, but that it should be understood as the view that cognition (human or otherwise) is computable (§2). On this view, it follows that computationalism can be true even if (human) cognition is not the result of computations in the brain. Second, I argue that, if semiotic systems are systems that interpret signs, then both humans and computers are semiotic systems (§5.2). Finally, I suggest that minds should be considered as virtual machines implemented in certain semiotic systems: primarily brains, but also AI computers (§6.3).
    [Show full text]
  • The Symbol Grounding Problem Has Been Solved, So What’S Next? Luc Steels
    12-de Vega-Chap12 5/7/08 6:26 PM Page 223 Chapter 12 The symbol grounding problem has been solved, so what’s next? Luc Steels 12.1 Introduction In the 1980s, a lot of ink was spent on the question of symbol grounding, largely triggered by Searle’s Chinese room theory (Searle 1980). Searle’s article had the advan- tage of stirring up discussion about when and how symbols could be about things in the world, whether intelligence involves representations or not, what embodiment means, and under what conditions cognition is embodied. But almost 25 years of philosophical discussion have shed little light on the issue, partly because the discussion has been mixed up with emotional arguments whether artificial intelligence (AI) is possible or not. However, today I believe that sufficient progress has been made in cognitive science and AI so that we can say that the symbol grounding problem has been solved. This chapter briefly discusses the issues of symbols, meanings, and embodiment (the main themes of the workshop), why I claim the symbol grounding problem has been solved, and what we should do next. 12.2 Symbols As suggested in Chapter 2, let us start from Peirce and the (much longer) semiotic tradi- tion which makes a distinction between a symbol, the objects in the world with which the symbol is associated (for example, for purposes of reference), and the concept associated with the symbol (see Figure 12.1). For example, we could have the symbol ‘ball’, a concrete round spherical object in the world with which a child is playing, and the concept ball which applies to this spherical object so that we can refer to the object using the symbol ‘ball’.
    [Show full text]
  • APA Newsletter on Philosophy and Computers, Vol. 16, No. 2
    APA NEWSLETTER | PHILOSOPHY AND COMPUTERS ———. “Philosophy of Computer Science: An Introductory Course.” ———. “Computational Theories of Cognition.” In The Philosophy Teaching Philosophy 28, no. 4 (2005b): 319–41. of Psychology, edited by W. O’Donohue and R. F. Kitchener, 160–72. London: SAGE Publications, 1996a. ———. “Semiotic Systems, Computers, and the Mind: How Cognition Could Be Computing.” International Journal of Signs and Semiotic ———. The Sciences of the Artificial, Third Edition. Cambridge, MA: The Systems 2, no. 1 (2012): 32–71. http://www.cse.buffalo.edu/~rapaport/ MIT Press, 1996b. Papers/Semiotic_Systems,_Computers,_and_the_Mind.pdf Sloman, A. The Computer Revolution in Philosophy: Philosophy, Science ———. “On the Relation of Computing to the World.” Forthcoming in and Models of Mind. Atlantic Highlands, NJ: Humanities Press, 1978. Philosophy and Computing: Essays in Epistemology, Philosophy of Mind, Logic, and Ethics, edited by T. M. Powers. Springer. Paper based on Smith, B.C. “Limits of Correctness in Computers.” ACM SIGCAS 2015 IACAP Covey Award talk; preprint available at http://www.cse. Computers and Society 14–15, nos. 1–4 (1985): 18–26. buffalo.edu/~rapaport/Papers/covey.pdf Soare, R. I. “Turing Oracle Machines, Online Computing, and Three ———. Philosophy of Computer Science. 2017. Available at Displacements in Computability Theory.” Annals of Pure and Applied http://www.cse.buffalo.edu/~rapaport/Papers/phics.pdf Logic 160 (2009): 368–99. Rapaport, W. J., and M. W. Kibby. “Contextual Vocabulary Acquisition as Srihari, S. N. “Beyond C.S.I.: The Rise of Computational Forensics.” Computational Philosophy and as Philosophical Computation.” Journal IEEE Spectrum. 2010. http://spectrum.ieee.org/computing/software/ of Experimental and Theoretical Artificial Intelligence 19, no.
    [Show full text]