University of Alberta

Total Page:16

File Type:pdf, Size:1020Kb

University of Alberta University of Alberta Istvh Stephen Norman Berkeley O A thesis submitted to the Faculty of Graduate Studies and Research in partial fulfillment of the requirements for the degree of Doctor of Philosophy Department of Philosophy Edmonton, Alberta Spring 1997 The author has granted a non- L'auteur a accordé une licence non exclusive iicence allowing the excImhe pumettant à la National Lr'brary of Canada to Bibliothèque nationale du Canada de reproduce, loan, districbute or selI nprodirm, prtkr, distnIbuerou copies of hismcr thesis by any means vmdn des copies de sa thèse de and in any fonn or fommt, maliiog 9ue1que manière et sow quelque this thesis adable to interested forme que ce soit pour mettre des persons. exemplaires de cette thèse a la disposition des personnes intéressées. The author retains ownership of the L'auteur conseme la propriété du copyrightin hislhtr thesis. Neither droit d'auteur qui protège sa thèse. Ni the thesis nor substantiai extracts la th& ni &s extraitssubstantieis de fiom it may be printed or othexwïse celle-ci ne doivent être imprimés ou reproduced with the author's autrement reproduits sans son permission. autorisation. Absttact This dissertation opens with a discussion and clarification of the Classical Computatioaal Theory of Miad (CCTM). An alleged challenge to this theory, which derives fiom research into connectionist systems, is then briefly outIined and connectionist systems are introduced in some detail. Several claims which have been made on behaif of connectiooist systems are then examined, these purport to support the conclusion that connectionist systems provide better models of the mind and cognitive functioning than the CCTM. It is argued that most claims made on behalf of connectionism often amount to litde more than myths. A significant dificuity with standard comectionist research methodology is then describeci. This dificuity derives hmthe fact that connectionist systems are seldom subject to detailed analysis and hterpretabion der tcaining and has serious consequemes for the plausibility of connectionist research in cognitive science and as the bais for a challenge to the CCTM. A technique for network anaiysis is described in detail and the results brn the anaiysis of a particuiar network are then introduced, in order to show how these difficdties can be overcome. The analyzed network exhiibited a number of surpnsing and intriguing propdes. These properties provide the basis for a detailed assessrnent of the network with respect to the CCTM. There then follows a discussion of the results of the network analysis, with respect to two feahues which are commonly associated with the CCTM, namely systematicity and compositionality. It is argued that the network has some properties similar to those associated with the CCTM and exhibits, in a weak seme, compositionality and systematicity. It is aiso argued that the network amounts to a genuinely cognitive theory. This suggests that there is insufncient evidence at the current tirne to determine whether or not comectionism presents a genuine challenge to the CCTM. The plausibility of the clai.that networks pmvide the basis of an alternative to the CCTM is then traced, in part, to a revisionkt tendency in some contempomy histories of comectionist research. Findy, fiture research directions in phiiosophy and cognitive science are suggested. 1 wouid like to thank the many pemple who assisted me with this dissertation. First and foremost, 1 would like to thank my Supavisors. Jeff Pelletier time and thne again went well above and beyond the cdof duty on di bnts, in addition to pmviding insightnil guidance and encouragement. Mïke Dawson (in conjunction with the various members of his Bio1ogica.i Computational Roject Iab) taught me how to be a cognitive scientist and provideci much of the inspiration for the empiricai work (and insights into it) dacniin this dissertation, as weil as providing the necessary facilties. 1 wouid also lüre to thank my examination committee, Wes Cooper, Robert Hadley, Bruce Hunter, Mohan Matthen and Gary Prideaux for thek many helpfûi commentS. In addition, 1 must thank the many people who gave me th& time and insight whilst 1 was writing this dissertation. There are too many to Est, but Ken Aizawa, Bill Bechtel, Andy Brook, Mason Cash, Andy Clark, Jerry Feldmau, Jeff Foss, Steve Giambroney Charles Ling, Tony Maida, Dave Medlery Ruth Michaels, Mawh MiaskyyZenon Pylyshyn, Don Ross, David Rumeihart, Paul Smolensky and Chris Thomton aii deserve special mention. AU the errors which remain are ail my own work. Table of Contents 1. Intniduction 1 IL The Classicd Cornputatbaai Tbeory of Mind (CCTM) 7 ID. Mat1s Coaaectionbm? 26 Introduction 26 Philosophical Objections to The Classicai Computatiod Theory of Mind 26 What 1s Co~mectionism? ProceSSitlg Units Leaming Rules Conclusion IV. The Myths of Coanectionism Introduction Myth 1 :Connectionist systems me biologikally plausible Processing Units Connections Myth 2: Connectionisl systems me consktent with mltime comtraints vonprocessing Myth 3: Connectionist Systems exhibit gracefl degradation Myîh 4: Connectionist systems are good generalizers Myth 5: Recent connectionist systenr have shown that Minsky und Papert were wrong Conclusion V. An Empirical Study: The Interpretation of The Logic Network, LI0 Introduction Ruies, Representations and Comectionist Systems McCloskey's Critique of the Comectionist Raearch Program Understanding Trained Comectionist Networks Bechtel and Abrahamsen's Logic Problem The Network L 10 The hoblem Encoding Scheme of LI0 The Training of LlO The Network Analysis Technique The Analysis of Network LI0 DefZnite Features and Inference 'Rules' Conclusion VI. Connectionist Networks and The CCTM Introduction The CCTM Again S ystematicity Compositionalîty Ruies and Cognitive Systems Conclusion VII. Conclusion: Connectionhm, Prcsent, Past and L;btPre Introduction: The Present History: The Past Furtber Research Directions: The Future Cognitive Science Philosophy List of Tables Table 2-1: Example of a Turing Machine Table 12 Table 5-1 :Examples of Valid Infetences hmBechtel and Abrabunsen's (1991) Logic Problem Set. Table 52: The Unary Definite Features Found in Band B of Hïdden Unit 4 of Network LI0 Table 13: The Binary Dennite Features Found in Band B of Hidden Unit 4 of Network LI0 Table 5-4: Interpretations of the Bands of the Hidden Units of Network LI0 97-98 Table 5-5: The Patterns of Bands Produced by LI0 for each Pmblem Type and the Properties Associated with each Band, as compared to the Properties Associated with the Inférence Rules of Natural Deduction Table 61: Network Rule for Modus Ponens, as Compared to the Traditional Rule Table 62: Network Rule for Type (i) Aitemative Syllogism, as Compared to the Traditional Rule List of Fimires Figure 2-1: Diagram of a Turing Machine Figure 2-2: Example of a Maze Figure 3-1: Detail of a Connectionist RocdgUnit Figure 3-2: A Step Activation Function Figure 3-3: A Sigrnoidal Activation Function Figure 3-4: A Gaussian Activation Function Figure 3-5: Detail of a Connection Weight Figure 3-6: Layen of Connections and Weights in a Connectionist Network Figure 5-1: The Structure and Encoding Scheme of the Network LI0 Figure 5-2: An Example of a Jittered Density Plot Figun 5-3: An Example Jittered Density Plot of a Hidden Value Unit Figmre 54: Jittered Density Plots for the 10 Hidden Units of Network LI0 Figure 5-5: The Jittered Density Plot of Hidden Unit 4 of Network L 10 1 introduction Issues arising hmthe computationai view ofthe mind will be central to this dissertation. There is a long tradition of philosophem taking inspiration for th& philosophicai theorizing hmmechanical devices. This was as tme at the time of Descartes, as it is toâay (see Meyering 1989). One particular contemporary instance of this tendency is the consideration of computational systems as a stnicturing metaphor for the mind (see Boden 1981). This strucnuing metaphor has inspimi an ambitious research program in cognitive science which, in turn, has inspird a considerable body of philosophical theorking about the mind. Central to both avenues of research are questions concerning representatiom or tokens. These questions arise as a direct consequeme of the adoption of the computationai metaphor. There are certain presumptions that have been made about the nature of representation or tokens, which have been centrai to both the philosophicai and tecbnical research programs. These presumptions define what has been called (e-g. by Fodor and Pylyshyn 1988) 'classical' cognitive science and philosophical theorizing about the mind. However, recently these suppositions have corne under attack. Advocates of the 'comectionist' approach have claimed that these suppositions are incorrect. Although it is usually assumed that there is some kind of incompatibility between the connectionist and classical positions, I will argue in this dissertation that these incompatibilities are not as great as has been assumed. In the chapter which immediately follows this one, introduce a paradigm exampie of a computational device, the Turing machine. Turing machines have played a signifîcant role in philosophical theorizing about the mind, especidy with respect to fùnctionalism. This Turing machine-like conception of the mind has (in conjmction with other factors, such as Fodor's (1975) language of thought hypothesis) underWntten a view of the mind, which I term the 'classical computationai theory of mind' (CCTM), which is central to much of both contemporary philosophy of mind and cognitive science. In this chapter 1 atternpt to consider in detail the properties wbich are supposeci by the CCTM to be shared by both Turing machina and minds. The purpose here is to develop a clear picture of the position, so that the alleged challenge to the CCTM posed by connectionist systems can be assesseci. Although the CCTM is weU established in the philosophical literature, it tas not been without its critics. Philosophers, such as Searle (1980), and Dreyfh (1991), have argued on a number of grounds that this position is deeply flawed.
Recommended publications
  • Generative Linguistics and Neural Networks at 60: Foundation, Friction, and Fusion*
    Generative linguistics and neural networks at 60: foundation, friction, and fusion* Joe Pater, University of Massachusetts Amherst October 3, 2018. Abstract. The birthdate of both generative linguistics and neural networks can be taken as 1957, the year of the publication of foundational work by both Noam Chomsky and Frank Rosenblatt. This paper traces the development of these two approaches to cognitive science, from their largely autonomous early development in their first thirty years, through their collision in the 1980s around the past tense debate (Rumelhart and McClelland 1986, Pinker and Prince 1988), and their integration in much subsequent work up to the present. Although this integration has produced a considerable body of results, the continued general gulf between these two lines of research is likely impeding progress in both: on learning in generative linguistics, and on the representation of language in neural modeling. The paper concludes with a brief argument that generative linguistics is unlikely to fulfill its promise of accounting for language learning if it continues to maintain its distance from neural and statistical approaches to learning. 1. Introduction At the beginning of 1957, two men nearing their 29th birthdays published work that laid the foundation for two radically different approaches to cognitive science. One of these men, Noam Chomsky, continues to contribute sixty years later to the field that he founded, generative linguistics. The book he published in 1957, Syntactic Structures, has been ranked as the most influential work in cognitive science from the 20th century.1 The other one, Frank Rosenblatt, had by the late 1960s largely moved on from his research on perceptrons – now called neural networks – and died tragically young in 1971.
    [Show full text]
  • History and Philosophy of Neural Networks
    HISTORY AND PHILOSOPHY OF NEURAL NETWORKS J. MARK BISHOP Abstract. This chapter conceives the history of neural networks emerging from two millennia of attempts to rationalise and formalise the operation of mind. It begins with a brief review of early classical conceptions of the soul, seating the mind in the heart; then discusses the subsequent Cartesian split of mind and body, before moving to analyse in more depth the twentieth century hegemony identifying mind with brain; the identity that gave birth to the formal abstractions of brain and intelligence we know as `neural networks'. The chapter concludes by analysing this identity - of intelligence and mind with mere abstractions of neural behaviour - by reviewing various philosophical critiques of formal connectionist explanations of `human understanding', `mathematical insight' and `consciousness'; critiques which, if correct, in an echo of Aristotelian insight, sug- gest that cognition may be more profitably understood not just as a result of [mere abstractions of] neural firings, but as a consequence of real, embodied neural behaviour, emerging in a brain, seated in a body, embedded in a culture and rooted in our world; the so called 4Es approach to cognitive science: the Embodied, Embedded, Enactive, and Ecological conceptions of mind. Contents 1. Introduction: the body and the brain 2 2. First steps towards modelling the brain 9 3. Learning: the optimisation of network structure 15 4. The fall and rise of connectionism 18 5. Hopfield networks 23 6. The `adaptive resonance theory' classifier 25 7. The Kohonen `feature-map' 29 8. The multi-layer perceptron 32 9. Radial basis function networks 34 10.
    [Show full text]
  • CAP 5636 - Advanced Artificial Intelligence
    CAP 5636 - Advanced Artificial Intelligence Introduction This slide-deck is adapted from the one used by Chelsea Finn at CS221 at Stanford. CAP 5636 Instructor: Lotzi Bölöni http://www.cs.ucf.edu/~lboloni/ Slides, homeworks, links etc: http://www.cs.ucf.edu/~lboloni/Teaching/CAP5636_Fall2021/index.html Class hours: Tue, Th 12:00PM - 1:15PM COVID considerations: UCF expects you to get vaccinated and wear a mask Classes will be in-person, but will be recorded on Zoom. Office hours will be over Zoom. Motivating artificial intelligence It is generally not hard to motivate AI these days. There have been some substantial success stories. A lot of the triumphs have been in games, such as Jeopardy! (IBM Watson, 2011), Go (DeepMind’s AlphaGo, 2016), Dota 2 (OpenAI, 2019), Poker (CMU and Facebook, 2019). On non-game tasks, we also have systems that achieve strong performance on reading comprehension, speech recognition, face recognition, and medical imaging benchmarks. Unlike games, however, where the game is the full problem, good performance on a benchmark does not necessarily translate to good performance on the actual task in the wild. Just because you ace an exam doesn’t necessarily mean you have perfect understanding or know how to apply that knowledge to real problems. So, while promising, not all of these results translate to real-world applications Dangers of AI From the non-scientific community, we also see speculation about the future: that it will bring about sweeping societal change due to automation, resulting in massive job loss, not unlike the industrial revolution, or that AI could even surpass human-level intelligence and seek to take control.
    [Show full text]
  • A Review of "Perceptrons: an Introduction to Computational Geometry"
    INFORMATION AND CONTROL 17, 501-522 (1970) A Review of "Perceptrons: An Introduction to Computational Geometry" by Marvin Minsky and Seymour Papert. The M.I.T. Press, Cambridge, Mass., 1969. 112 pages. Price: Hardcover $12.00; Paperback $4.95. H. D. BLOCK Department of Theoretical and Applied 2Plechanics, Cornell University, Ithaca, New York 14850 1. INTRODUCTION The purpose of this book is to present a mathematical theory of the class of machines known as Perceptrons. The theory is carefully formulated and focuses on the theoretical capabilities and limitations of these machines. It is a remarkable book. Not only do the authors formulate a new and fundamental conceptual framework, but they also fill in the details using strikingly ingenious mathematical techniques. They ask some novel questions and find some difficult answers. The most striking of these will be presented in Section 2. The authors address the book to three classes of readers: (1) Computer scientists, specializing in pattern recognition, learning machines, and threshold logic; (2) Abstract mathematicians interested in the d6but of Computational Geometry; (3) Those interested in a general theory of computation leading to decisions based on the weight of partial evidence. The authors hope that this class includes psychologists and biologists. In Section 6 I shall give my estimate of the value of the book to each of these groups. The conversational style and the childlike freehand sketches might mislead the casual reader into believing that this book makes light reading. For example, the review in The American Mathematical Monthly (1969) states that the prospective reader "requires little mathematics beyond the high 501 502 BLOCK school level." This, as we shall see, is somewhat sanguine.
    [Show full text]
  • A Revisionist History of Connectionism
    History: The Past Seite 1 von 5 Copyright 1997, All rights reserved. A Revisionist History of Connectionism Istvan S. N. Berkeley Bibliography According to the standard (recent) history of connectionism (see for example the accounts offered by Hecht-Nielsen (1990: pp. 14-19) and Dreyfus and Dreyfus (1988), or Papert's (1988: pp. 3-4) somewhat whimsical description), in the early days of Classical Computational Theory of Mind (CCTM) based AI research, there was also another allegedly distinct approach, one based upon network models. The work on network models seems to fall broadly within the scope of the term 'connectionist' (see Aizawa 1992), although the term had yet to be coined at the time. These two approaches were "two daughter sciences" according to Papert (1988: p. 3). The fundamental difference between these two 'daughters', lay (according to Dreyfus and Dreyfus (1988: p. 16)) in what they took to be the paradigm of intelligence. Whereas the early connectionists took learning to be fundamental, the traditional school concentrated upon problem solving. Although research on network models initially flourished along side research inspired by the CCTM, network research fell into a rapid decline in the late 1960's. Minsky (aided and abetted by Papert) is often credited with having personally precipitated the demise of research in network models, which marked the end of the first phase of connectionist research. Hecht-Nielson (1990: pp. 16-17) describes the situation (as it is presented in standard versions of the early history of connectionism)
    [Show full text]
  • AI Patents: a Data Driven Approach
    Chicago-Kent Journal of Intellectual Property Volume 19 Issue 3 Article 6 6-25-2020 AI Patents: A Data Driven Approach Brian S. Haney Notre Dame Law School, [email protected] Follow this and additional works at: https://scholarship.kentlaw.iit.edu/ckjip Part of the Intellectual Property Law Commons Recommended Citation Brian S. Haney, AI Patents: A Data Driven Approach, 19 Chi. -Kent J. Intell. Prop. 407 (2020). Available at: https://scholarship.kentlaw.iit.edu/ckjip/vol19/iss3/6 This Article is brought to you for free and open access by Scholarly Commons @ IIT Chicago-Kent College of Law. It has been accepted for inclusion in Chicago-Kent Journal of Intellectual Property by an authorized editor of Scholarly Commons @ IIT Chicago-Kent College of Law. For more information, please contact [email protected], [email protected]. AI Patents: A Data Driven Approach Cover Page Footnote Thanks to Angela Elias, Dean Alderucci, Tabrez Y. Ebrahim, Mark Opitz, Barry Coyne, Brian Bozzo, Tom Sweeney, Richard Susskind, Ron Dolin, Mike Gallagher, Branden Keck. This article is available in Chicago-Kent Journal of Intellectual Property: https://scholarship.kentlaw.iit.edu/ckjip/ vol19/iss3/6 AI PATENTS: A DATA DRIVEN APPROACH 5/29/2020 6:58 PM AI PATENTS: A DATA DRIVEN APPROACH BRIAN S. HANEY* Contents I. Introduction ........................................................................................... 410 A. Definition ........................................................................................... 410 B. Applications .....................................................................................
    [Show full text]
  • Scale, Abstraction, and Connectionist Models: on Parafinite Thresholds in Artificial Intelligence
    Scale, abstraction, and connectionist models: On parafinite thresholds in Artificial Intelligence by Zachary Peck (Under the Direction of O. Bradley Bassler) Abstract In this thesis, I investigate the conceptual intersection between scale, abstraction, and connectionist modeling in articial intelligence. First, I argue that connectionist learning algorithms allow for higher levels of abstraction to dynamically emerge when paranite thresholds are surpassed within the network. Next, I argue that such networks may be surveyable, provided we evaluate them at the appropriate level of abstraction. To demonstrate, I consider the recently released GPT-3 as a semantic model in contrast to logical semantic models in the logical inferentialist tradition. Finally, I argue that connectionist models capture the indenite nature of higher-level abstractions, such as those that appear in natural language. Index words: [Connectionism, Scale, Levels, Abstraction, Neural Networks, GPT-3] Scale, abstraction, and connectionist models: On parafinite thresholds in Artificial Intelligence by Zachary Peck B.S., East Tennessee State University, 2015 M.A., Georgia State University, 2018 A Thesis Submitted to the Graduate Faculty of the University of Georgia in Partial Fulllment of the Requirements for the Degree. Master of Science Athens, Georgia 2021 ©2021 Zachary Peck All Rights Reserved Scale, abstraction, and connectionist models: On parafinite thresholds in Artificial Intelligence by Zachary Peck Major Professor: O. Bradley Bassler Committee: Frederick Maier Robert Schneider Electronic Version Approved: Ron Walcott Dean of the Graduate School The University of Georgia May 2021 Contents List of Figures v 1 Introduction 1 2 From Perceptrons to GPT-3 5 2.1 Perceptron . 5 2.2 Multi-layer, fully connected neural nets .
    [Show full text]
  • Arxiv:2104.12871V2 [Cs.AI] 28 Apr 2021 H Er22 a Upsdt Eadtearvlo Efdiigca Self-Driving of Arrival the Herald to Supposed Guardian Was 2020 Year the Introduction Sense
    Why AI is Harder Than We Think Melanie Mitchell Santa Fe Institute Santa Fe, NM, USA [email protected] Abstract Since its beginning in the 1950s, the field of artificial intelligence has cycled several times between periods of optimistic predictions and massive investment (“AI spring”) and periods of disappointment, loss of confi- dence, and reduced funding (“AI winter”). Even with today’s seemingly fast pace of AI breakthroughs, the development of long-promised technologies such as self-driving cars, housekeeping robots, and conversational companions has turned out to be much harder than many people expected. One reason for these repeating cycles is our limited understanding of the nature and complexity of intelligence itself. In this paper I describe four fallacies in common assumptions made by AI researchers, which can lead to overconfident predictions about the field. I conclude by discussing the open questions spurred by these fallacies, including the age-old challenge of imbuing machines with humanlike common sense. Introduction The year 2020 was supposed to herald the arrival of self-driving cars. Five years earlier, a headline in The Guardian predicted that “From 2020 you will become a permanent backseat driver” [1]. In 2016 Business Insider assured us that “10 million self-driving cars will be on the road by 2020” [2]. Tesla Motors CEO Elon Musk promised in 2019 that “A year from now, we’ll have over a million cars with full self-driving, software...everything” [3]. And 2020 was the target announced by several automobile companies to bring self-driving cars to market [4, 5, 6].
    [Show full text]
  • Perceptrons: an Associative Learning Network
    Perceptrons: An Associative Learning Network by Michele D. Estebon Virginia Tech CS 3604, Spring 1997 In the early days of artificial intelligence research, Frank Rosenblatt devised a machine called the perceptron that operated much in the same way as the human mind. Albeit, it did not have what could be called a "mental capacity", it could "learn" - and that was the breakthrough needed to pioneer todayís current neural network technologies. A perceptron is a connected network that simulates an associative memory. The most basic perceptron is composed of an input layer and output layer of nodes, each of which are fully connected to the other. Assigned to each connection is a weight which can be adjusted so that, given a set of inputs to the network, the associated connections will produce a desired output. The adjusting of weights to produce a particular output is called the "training" of the network which is the mechanism that allows the network to learn. Perceptrons are among the earliest and most basic models of artificial neural networks, yet they are at work in many of todayís complex neural net applications. An Associative Network Model In order to understand the operations of the human brain, many researchers have chosen to create models based on findings in biological neural research. Neurobiologists have shown that human nerve cells have no memory of their own, rather memory is made up of a series of connections, or associations, throughout the nervous system. A response in one cell triggers responses in others which triggers more in others until these responses eventually lead to a so-called ìrecognitionî of the input stimuli.
    [Show full text]
  • A Brief History of Deep Learning
    A brief history of deep learning Alice Gao Lecture 10 Readings: this article by Andrew Kurenkov Based on work by K. Leyton-Brown, K. Larson, and P. van Beek 1/11 Outline Learning Goals A brief history of deep learning Lessons learned summarized by Geoff Hinton Revisiting the Learning goals 2/11 Learning Goals By the end of the lecture, you should be able to I Describe the causes and the resolutions of the two AI winters in the past. I Describe the four lessons learned summarized by Geoff Hinton. 3/11 A brief history of deep learning A brief history of deep learning based on this article by Andrew Kurenkov. 4/11 The birth of machine learning (Frank Rosenblatt 1957) I A perceptron can be used to represent and learn logic operators (AND, OR, and NOT). I It was widely believed that AI is solved if computers can perform formal logical reasoning. 5/11 First AI winter (Marvin Minsky 1969) I A rigorous analysis of the limitations of perceptrons. I (1) We need to use multi-layer perceptrons to represent simple non-linear functions such as XOR. I (2) No one knows a good way to train multi-layer perceptrons. I Led to the first AI winter from the 70s to the 80s. 6/11 The back-propagation algorithm (Rumelhart, Hinton and Williams 1986) I In their Nature article, they precisely and concisely explained how we can use the back-propagation algorithm to train multi-layer perceptrons. 7/11 Second AI Winter I Multi-layer neural networks trained with back-propagation don’t work well, especially compared to simpler models (e.g.
    [Show full text]
  • The Next Decade in AI: Four Steps Towards Robust Artificial Intelligence
    The Next Decade in AI: Four Steps Towards Robust Artificial Intelligence Gary Marcus Robust AI 17 February 2020 Abstract Recent research in artificial intelligence and machine learning has largely emphasized general-purpose learning and ever-larger training sets and more and more compute. In contrast, I propose a hybrid, knowledge-driven, reasoning-based approach, centered around cognitive models, that could provide the substrate for a richer, more robust AI than is currently possible. THE NEXT DECADE IN AI / GARY MARCUS Table of Contents 1. Towards robust artificial intelligence 3 2. A hybrid, knowledge-driven, cognitive-model-based approach 8 2.1. Hybrid architecture 14 2.2. Large-scale knowledge, some of which is abstract and causal 24 2.3. Reasoning 36 2.4. Cognitive models 40 3. Discussion 47 3.1. Towards an intelligence framed around enduring, abstract knowledge 47 3.2 Is there anything else we can do? 51 3.3. Seeing the whole elephant, a little bit at a time 52 3.4 Conclusions, prospects, and implications 53 4. Acknowledgements 55 5. References 55 2 THE NEXT DECADE IN AI / GARY MARCUS [the] capacity to be affected by objects, must necessarily precede all intuitions of these objects,.. and so exist[s] in the mind a priori. — Immanuel Kant Thought is ... a kind of algebra … — William James You can’t model the probability distribution function for the whole world, because the world is too complicated. — Eero Simoncelli 1. Towards robust artificial intelligence Although nobody quite knows what deep learning or AI will evolve into the coming decades, it is worth considering both what has been learned from the last decade, and what should be investigated next, if we are to reach a new level.
    [Show full text]
  • 1956 and the Origins of Artificial Intelligence Computing
    Building the Second Mind: 1956 and the Origins of Artificial Intelligence Computing Rebecca E. Skinner copyright 2012 by Rebecca E. Skinner ISBN 978-0-9894543-1-5 Building the Second Mind: 1956 and the Origins of Artificial Intelligence Computing Chapter .5. Preface Chapter 1. Introduction Chapter 2. The Ether of Ideas in the Thirties and the War Years Chapter 3. The New World and the New Generation in the Forties Chapter 4. The Practice of Physical Automata Chapter 5. Von Neumann, Turing, and Abstract Automata Chapter 6. Chess, Checkers, and Games Chapter 7. The Impasse at End of Cybernetic Interlude Chapter 8. Newell, Shaw, and Simon at the Rand Corporation Chapter 9. The Declaration of AI at the Dartmouth Conference Chapter 10. The Inexorable Path of Newell and Simon Chapter 11. McCarthy and Minsky begin Research at MIT Chapter 12. AI’s Detractors Before its Time Chapter 13. Conclusion: Another Pregnant Pause Chapter 14. Acknowledgements Chapter 15. Bibliography Chapter 16. Endnotes Chapter .5. Preface Introduction Building the Second Mind: 1956 and the Origins of Artificial Intelligence Computing is a history of the origins of AI. AI, the field that seeks to do things that would be considered intelligent if a human being did them, is a universal of human thought, developed over centuries. Various efforts to carry this out appear- in the forms of robotic machinery and more abstract tools and systems of symbols intended to artificially contrive knowledge. The latter sounds like alchemy, and in a sense it certainly is. There is no gold more precious than knowledge. That this is a constant historical dream, deeply rooted in the human experience, is not in doubt.
    [Show full text]