Analogy and Qualitative Representations in the Companion Cognitive Architecture
Total Page:16
File Type:pdf, Size:1020Kb
Articles Analogy and Qualitative Representations in the Companion Cognitive Architecture Kenneth D. Forbus, Thomas Hinrichs n The Companion cognitive architec - very cognitive architecture starts with a set of theoreti - ture is aimed at reaching human-level cal commitments. We have argued (Forbus 2016) that AI by creating software social organ - Ehuman-level artificial intelligences will be built by cre - isms — systems that interact with peo - ating sufficiently smart software social organisms. By that we ple using natural modalities, working mean systems capable of interacting with people using natu - and learning over extended periods of ral modalities, operating and learning over extended periods time as collaborators rather than tools. Our two central hypotheses about how of time, as apprentices and collaborators, instead of as tools. to achieve this are (1) analogical rea - Just as we cannot directly access the internal representations soning and learning are central to cog - of the people and animals we work with, cognitive systems nition, and (2) qualitative representa - should be able to work with us on our terms. But how does tions provide a level of description that one create such systems? We have two core hypotheses, facilitates reasoning, learning, and inspired by research in cognitive science: communication. This article discusses Our first core hypothesis is that analogical reasoning and the evidence we have gathered support - learning are central to human cognition. There is evidence that ing these hypotheses from our experi - processes described by Gentner’s (1983) structure-mapping ments with the Companion architec - ture. Although we are far from our theory of analogy and similarity operate throughout human ultimate goals, these experiments pro - cognition, including visual perception (Sagi, Gentner, and vide strong evidence for the utility of Lovett 2012), reasoning and decision making (Markman and analogy and qualitative representation Medin 2002), and conceptual change (Gentner et al. 1997). across a range of tasks. We also discuss Our second core hypothesis is that qualitative representations three lessons learned and highlight three (QRs) are a key building block of human conceptual structure. Con - important open problems for cognitive tinuous phenomena and systems permeate our environment systems research more broadly. and our ways of thinking about it. This includes the physical world, where qualitative representations have a long track record of providing human-level reasoning and performance (Forbus 2014), but also in social reasoning (for example, degrees of blame [Tomai and Forbus 2007]). Qualitative rep - resentations carve up continuous phenomena into symbolic descriptions that serve as a bridge between perception and cognition, facilitate everyday reasoning and communication, and help ground expert reasoning. 34 AI MAGAZINE Copyright © 2017, Association for the Advancement of Artificial Intelligence. All rights reserved. ISSN 0738-4602 Articles The focus of the Companion cognitive architec - roles in cognition. We close with some lessons ture (Forbus, Klenk, and Hinrichs 2009) is on high - learned and open problems. er-order cognition: conceptual reasoning and learn - ing, and learning through interactions with others, by contrast with architectures that have focused on Analogical Reasoning and Learning skill learning (for example, ACT-R [Anderson 2009] Dedre Gentner’s (1983) structure-mapping theory and SOAR [Laird 2012]). In Newell’s (1990) time- proposed that analogy involves the construction of scale decomposition of cognitive phenomena, con - mappings between two structured, relational repre - ceptual reasoning and learning occur in what are sentations. These mappings contain correspondences called the rational and social bands ,1 unlike many (that is, what goes with what), candidate inferences architectures, which start with Newell’s cognitive (that is, what can be projected from one description band (Laird, Lebiere, and Rosenbloom 2017). Thus to the other, based on the correspondences), and a we approximate subsystems whose operations occur score indicating the overall quality of the match. at faster time scales, using methods whose outputs Typically, the base is the description about which have reasonable cognitive fidelity, although we are more is known, and the target is the description about not making theoretical claims about them. For which one is trying to reason about, and hence infer - example, in Companions constraint checking and ences are made from base to target by default. Infer - simple inferences are carried out through a logic- ences in the other direction, reverse candidate infer - based truth-maintenance system (Forbus and de ences, can also be drawn, which is important for Kleer 1994). Similarly, the Companion architecture difference detection, discussed more later. is implemented as a multiagent system, capable of There is ample evidence for the psychological plau - running on a single laptop or distributed across sibility of structure mapping, out of Gentner’s lab many cluster nodes, depending on the task. We sus - and others (Forbus 2001). This evidence includes pect that many data-parallel operations and a num - some implications that may surprise AI researchers. ber of coarse-grained parallel operations are impor - Here are three such findings: tant for creating robust software organisms, but this (1) The same computations that people use for current organization is motivated more by engineer - analogy are also used for everyday similarity (Mark - ing concerns. By contrast, we spend considerable man and Gentner 1993). Most AI researchers model effort on sketch understanding and natural language similarity as the dot product of feature vectors or oth - understanding, since interaction through natural er distributed representations. But those approaches modalities is a core concern. are not compatible with psychological evidence that Cognitive systems are knowledge rich (for exam - indicates, even for visual stimuli, that relational rep - ple, McShane [2017]). The Companion architecture resentations are important (Goldstone, Medin, and uses the Cyc 2 ontology and knowledge base (KB) con - Gentner 1991). tents, plus our own extensions including additional (2) Differences are closely tied to similarity (Mark - linguistic resources, representations for visual/spatial man and Gentner 1996). There is an interesting dis - reasoning, and the hierarchical task network (HTN) sociation between difference detection and naming a plans that drive Companion operations. This is in difference: it is faster to detect that things are differ - sharp distinction with current practice in machine ent when they are very different, but faster to name learning, where the goal is that systems for every task a difference when they are very similar (Sagi, Gen - must be learned from scratch. The machine-learning tner, and Lovett 2012). This falls directly out of our (ML) approach can be fine for specific applications, computational model of analogical matching below. but it means that many independently learnable fac - (3) Analogical comparison, especially within- tors in a complex task must be learned at the same domain comparison, can happen unconsciously. For time. This intermingling leads to requiring far more example, people reading a story can be pushed training data than people need, and reduces transfer between two different interpretations of it, based on of learned knowledge from one task to another. The a different story that they read previously (Day and cognitive-systems approach strives for cumulative Gentner 2007). Moreover, they report being sure learning, where an architecture that is general pur - that the second story was complete in itself, and pose can learn to do one task and successfully applies nothing from the previous story influenced them. that knowledge to learn other tasks. This implies, for example, that in any of the System The remainder of this article summarizes the evi - 1 / System 2 models (for example, Kahneman dence provided by Companion experiments for our [2011]), analogy and similarity are actually used in core hypotheses. We start by summarizing the ideas both systems. underlying our hypotheses, briefly introducing struc - Our computational models of analogical processes ture mapping and our models of its processes for have been designed both as cognitive models and as analogy followed by the key aspects of qualitative performance systems. That is, each of them has been reasoning. Then we describe the evidence that anal - used to explain (and sometimes predict) psychologi - ogy and qualitative representations play a variety of cal findings, while also being used in AI systems WINTER 2017 35 Articles whose sole goal was achieving some new functional - world, we must deal with quantities like heat, tem - ity. These models are SME, MAC/FAC, and SAGE. perature, and mass, and arrange for processes like The structure-mapping engine (SME) (Forbus et al. flows and phase changes to cook. Spatial factors are 2016) models analogical mapping. It computes up to key in many physical situations, such as finding safe three mappings, using a greedy algorithm to operate ground during a flood. In social reasoning, we think in polynomial time. It can also operate incremental - about quantities like degree of blame, how strong a ly, extending mappings as new facts are added to the friendship is, and the relative morality of choices we base or target, which can