CAP 5636 - Advanced Artificial Intelligence
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Generative Linguistics and Neural Networks at 60: Foundation, Friction, and Fusion*
Generative linguistics and neural networks at 60: foundation, friction, and fusion* Joe Pater, University of Massachusetts Amherst October 3, 2018. Abstract. The birthdate of both generative linguistics and neural networks can be taken as 1957, the year of the publication of foundational work by both Noam Chomsky and Frank Rosenblatt. This paper traces the development of these two approaches to cognitive science, from their largely autonomous early development in their first thirty years, through their collision in the 1980s around the past tense debate (Rumelhart and McClelland 1986, Pinker and Prince 1988), and their integration in much subsequent work up to the present. Although this integration has produced a considerable body of results, the continued general gulf between these two lines of research is likely impeding progress in both: on learning in generative linguistics, and on the representation of language in neural modeling. The paper concludes with a brief argument that generative linguistics is unlikely to fulfill its promise of accounting for language learning if it continues to maintain its distance from neural and statistical approaches to learning. 1. Introduction At the beginning of 1957, two men nearing their 29th birthdays published work that laid the foundation for two radically different approaches to cognitive science. One of these men, Noam Chomsky, continues to contribute sixty years later to the field that he founded, generative linguistics. The book he published in 1957, Syntactic Structures, has been ranked as the most influential work in cognitive science from the 20th century.1 The other one, Frank Rosenblatt, had by the late 1960s largely moved on from his research on perceptrons – now called neural networks – and died tragically young in 1971. -
AI Thinks Like a Corporation—And That's Worrying
Topics Current edition More Subscribe Log in or sign up Search Open Future Manage subscription Open Voices AI thinks like a corporation—and that’s worrying Arti!cial intelligence was born of organisational decision-making and state power; it needs human ethics, says Jonnie Penn of the University of Cambridge Open Future Nov 26th 2018 | by BY JONNIE PENN Arti!cial intelligence is everywhere but it is considered in a wholly ahistorical way. To understand the impact AI will have on our lives, it is vital to appreciate the context in which the !eld was established. After all, statistics and state control have evolved hand in hand for hundreds of years. Consider computing. Its origins have been traced not only to analytic philosophy, pure mathematics and Alan Turing, but perhaps surprisingly, to the history of public administration. In “The Government Machine: A Revolutionary History of the Computer” from 2003, Jon Agar of University College London charts the development of the British civil service as it ballooned from 16,000 employees in 1797 to 460,000 by 1999. He noticed an uncanny similarity between the functionality of a human bureaucracy and that of the digital electronic computer. (He confessed that he could not tell whether this observation was trivial or profound.) Get our daily newsletter Upgrade your inbox and get our Daily Dispatch and Editor's Picks. Email address Sign up now Both systems processed large quantities of information using a hierarchy of pre-set but adaptable rules. Yet one predated the other. This suggested a telling link between the organisation of human social structures and the digital tools designed to serve them. -
History and Philosophy of Neural Networks
HISTORY AND PHILOSOPHY OF NEURAL NETWORKS J. MARK BISHOP Abstract. This chapter conceives the history of neural networks emerging from two millennia of attempts to rationalise and formalise the operation of mind. It begins with a brief review of early classical conceptions of the soul, seating the mind in the heart; then discusses the subsequent Cartesian split of mind and body, before moving to analyse in more depth the twentieth century hegemony identifying mind with brain; the identity that gave birth to the formal abstractions of brain and intelligence we know as `neural networks'. The chapter concludes by analysing this identity - of intelligence and mind with mere abstractions of neural behaviour - by reviewing various philosophical critiques of formal connectionist explanations of `human understanding', `mathematical insight' and `consciousness'; critiques which, if correct, in an echo of Aristotelian insight, sug- gest that cognition may be more profitably understood not just as a result of [mere abstractions of] neural firings, but as a consequence of real, embodied neural behaviour, emerging in a brain, seated in a body, embedded in a culture and rooted in our world; the so called 4Es approach to cognitive science: the Embodied, Embedded, Enactive, and Ecological conceptions of mind. Contents 1. Introduction: the body and the brain 2 2. First steps towards modelling the brain 9 3. Learning: the optimisation of network structure 15 4. The fall and rise of connectionism 18 5. Hopfield networks 23 6. The `adaptive resonance theory' classifier 25 7. The Kohonen `feature-map' 29 8. The multi-layer perceptron 32 9. Radial basis function networks 34 10. -
The Logicist Manifesto: at Long Last Let Logic-Based Artificial Intelligence Become a Field Unto Itself∗
The Logicist Manifesto: At Long Last Let Logic-Based Artificial Intelligence Become a Field Unto Itself∗ Selmer Bringsjord Rensselaer AI & Reasoning (RAIR) Lab Department of Cognitive Science Department of Computer Science Rensselaer Polytechnic Institute (RPI) Troy NY 12180 USA [email protected] version 9.18.08 Contents 1 Introduction 1 2 Background 1 2.1 Logic-Based AI Encapsulated . .1 2.1.1 LAI is Ambitious . .3 2.1.2 LAI is Based on Logical Systems . .4 2.1.3 LAI is a Top-Down Enterprise . .5 2.2 Ignoring the \Strong" vs. \Weak" Distinction . .5 2.3 A Slice in the Day of a Life of a LAI Agent . .6 2.3.1 Knowledge Representation in Elementary Logic . .8 2.3.2 Deductive Reasoning . .8 2.3.3 A Note on Nonmonotonic Reasoning . 12 2.3.4 Beyond Elementary Logical Systems . 13 2.4 Examples of Logic-Based Cognitive Systems . 15 3 Factors Supporting Logicist AI as an Independent Field 15 3.1 History Supports the Divorce . 15 3.2 The Advent of the Web . 16 3.3 The Remarkable Effectiveness of Logic . 16 3.4 Logic Top to Bottom Now Possible . 17 3.5 Learning and Denial . 19 3.6 Logic is an Antidote to \Cheating" in AI . 19 3.7 Logic Our Only Hope Against the Dark AI Future . 21 4 Objections; Rebuttals 22 4.1 \But you are trapped in a fundamental dilemma: your position is either redundant, or false." . 22 4.2 \But you're neglecting probabilistic AI." . 23 4.3 \But we now know that the mind, contra logicists, is continuous, and hence dynamical, not logical, systems are superior." . -
A Review of "Perceptrons: an Introduction to Computational Geometry"
INFORMATION AND CONTROL 17, 501-522 (1970) A Review of "Perceptrons: An Introduction to Computational Geometry" by Marvin Minsky and Seymour Papert. The M.I.T. Press, Cambridge, Mass., 1969. 112 pages. Price: Hardcover $12.00; Paperback $4.95. H. D. BLOCK Department of Theoretical and Applied 2Plechanics, Cornell University, Ithaca, New York 14850 1. INTRODUCTION The purpose of this book is to present a mathematical theory of the class of machines known as Perceptrons. The theory is carefully formulated and focuses on the theoretical capabilities and limitations of these machines. It is a remarkable book. Not only do the authors formulate a new and fundamental conceptual framework, but they also fill in the details using strikingly ingenious mathematical techniques. They ask some novel questions and find some difficult answers. The most striking of these will be presented in Section 2. The authors address the book to three classes of readers: (1) Computer scientists, specializing in pattern recognition, learning machines, and threshold logic; (2) Abstract mathematicians interested in the d6but of Computational Geometry; (3) Those interested in a general theory of computation leading to decisions based on the weight of partial evidence. The authors hope that this class includes psychologists and biologists. In Section 6 I shall give my estimate of the value of the book to each of these groups. The conversational style and the childlike freehand sketches might mislead the casual reader into believing that this book makes light reading. For example, the review in The American Mathematical Monthly (1969) states that the prospective reader "requires little mathematics beyond the high 501 502 BLOCK school level." This, as we shall see, is somewhat sanguine. -
Frameworks for Intelligent Systems
Frameworks for Intelligent Systems CompSci 765 Meeting 3 Pat Langley Department of Computer Science University of Auckland Outline of the Lecture • Computer science as an empirical discipline • Physical symbol systems • List structures and list processing • Reasoning and intelligence • Intelligence and search • Knowledge and intelligence • Implications for social cognition 2 Computer Science as an Empirical Discipline! In their Turing Award article, Newell and Simon (1976) make some important claims: • Computer science is an empirical discipline, rather than a branch of mathematics. • It is a science of the artificial, in that it constructs artifacts of sufficient complexity that formal analysis is not tractable. • We must study these computational artifacts as if they were natural systems, forming hypotheses and collecting evidence. They propose two hypotheses based on their founding work in list processing and artificial intelligence.! 3 Laws of Qualitative Structure! The authors introduce the idea of laws of qualitative structure, which are crucial for any scientific field’s development: • The cell doctrine in biology • Plate tectonics in geology • The germ theory of disease • The atomic theory of matter They propose two such laws, one related to mental structures and another and the other to mental processes. 4 Physical Symbol Systems! Newell and Simon’s first claim, the physical symbol system hypothesis, states that: • A physical symbol system has the necessary and sufficient means for general intelligent action. They emphasize general cognitive abilities, such as humans exhibit, rather than specialized ones. This is a theoretical claim that is subject to empirical tests, but the evidence to date generally supports it.! 5 More on Physical Symbol Systems! What do Newell and Simon mean by a physical symbol system? • Symbols are physical patterns that are stable unless modified. -
After Math: (Re)Configuring Minds, Proof, and Computing in the Postwar United States
After Math: (Re)configuring Minds, Proof, and Computing in the Postwar United States The Harvard community has made this article openly available. Please share how this access benefits you. Your story matters Citation Dick, Stephanie Aleen. 2015. After Math: (Re)configuring Minds, Proof, and Computing in the Postwar United States. Doctoral dissertation, Harvard University, Graduate School of Arts & Sciences. Citable link http://nrs.harvard.edu/urn-3:HUL.InstRepos:14226096 Terms of Use This article was downloaded from Harvard University’s DASH repository, and is made available under the terms and conditions applicable to Other Posted Material, as set forth at http:// nrs.harvard.edu/urn-3:HUL.InstRepos:dash.current.terms-of- use#LAA After Math (Re)configuring Minds, Proof, and Computing in the Postwar United States Adissertationpresented by Stephanie Aleen Dick to The Department of the History of Science in partial fulfillment of the requirements for the degree of Doctor of Philosophy in the subject of the History of Science Harvard University Cambridge, Massachusetts November 2014 © 2014 Stephanie Aleen Dick. All rights reserved. Dissertation Advisor: Professor Peter Galison Stephanie Aleen Dick After Math (Re)configuring Minds, Proof, and Computing in the Postwar United States Abstract This dissertation examines the history of three early computer programs that were designed to prove mathematical theorems: The Logic Theory Machine, the Program P, and the Automated Reasoning Assistant, all developed between 1955 and 1975. I use these programs as an opportunity to explore ways in which mathematical knowledge and practice were transformed by the introduction of modern computing. The prospect of automation generated disagreement about the character of human mathematical faculties like intuition, reasoning, and understanding and whether computers could be made to possess or participate in them. -
OCAMS Methods AIEDAM V2
Workflow Agents vs. Expert Systems: Problem Solving Methods in Work Systems Design William J. Clancey NASA Ames Research Center & Florida Institute for Human & Machine Cognition Maarten Sierhuis RIACS, NASA Ames Research Center Chin Seah QSS, NASA Ames Research Center Prepared for special issue of Artificial Intelligence for Engineering Design, Analysis, and Manufacturing (AIEDAM) — “Problem Solving Methods: Past, Present, and Future,” Spring 2008 Corresponding Author: [email protected] Intelligent Systems Division, M/S 269-3 NASA Ames Research Center Moffett Field, CA 94035 650-604-2526, fax 4-4036 Short Title: “The Role of Method Abstraction in Work Systems Design” Number of pages (excluding title, abstract, and biographies): 55 Number of figures: 2 Draft: May 1, 2008 1 Workflow Agents vs. Expert Systems: Problem Solving Methods in Work Systems Design Abstract During the 1980s, a community of artificial intelligence researchers became interested in formalizing problem solving methods as part of an effort called “second generation expert systems” (2nd GES). How do the motivations and results of this research relate to building tools for the workplace today? We provide an historical review of how the theory of expertise has developed, a progress report on a tool for designing and implementing model-based automation (Brahms), and a concrete example how we apply 2nd GES concepts today in an agent-based system for space flight operations (OCAMS). Brahms’ incorporates an ontology for modeling work practices, what people are doing in the course of a day, characterized as “activities.” OCAMS was developed using a simulation-to-implementation methodology, in which a prototype tool was embedded in a simulation of future work practices. -
University of Alberta
University of Alberta Istvh Stephen Norman Berkeley O A thesis submitted to the Faculty of Graduate Studies and Research in partial fulfillment of the requirements for the degree of Doctor of Philosophy Department of Philosophy Edmonton, Alberta Spring 1997 The author has granted a non- L'auteur a accordé une licence non exclusive iicence allowing the excImhe pumettant à la National Lr'brary of Canada to Bibliothèque nationale du Canada de reproduce, loan, districbute or selI nprodirm, prtkr, distnIbuerou copies of hismcr thesis by any means vmdn des copies de sa thèse de and in any fonn or fommt, maliiog 9ue1que manière et sow quelque this thesis adable to interested forme que ce soit pour mettre des persons. exemplaires de cette thèse a la disposition des personnes intéressées. The author retains ownership of the L'auteur conseme la propriété du copyrightin hislhtr thesis. Neither droit d'auteur qui protège sa thèse. Ni the thesis nor substantiai extracts la th& ni &s extraitssubstantieis de fiom it may be printed or othexwïse celle-ci ne doivent être imprimés ou reproduced with the author's autrement reproduits sans son permission. autorisation. Absttact This dissertation opens with a discussion and clarification of the Classical Computatioaal Theory of Miad (CCTM). An alleged challenge to this theory, which derives fiom research into connectionist systems, is then briefly outIined and connectionist systems are introduced in some detail. Several claims which have been made on behaif of connectiooist systems are then examined, these purport to support the conclusion that connectionist systems provide better models of the mind and cognitive functioning than the CCTM. -
ALLEN NEWELL March 19, 1927–July 19, 1992
NATIONAL ACADEMY OF SCIENCES A L L E N N E W ELL 1927—1992 A Biographical Memoir by HER BE R T A. S IMON Any opinions expressed in this memoir are those of the author(s) and do not necessarily reflect the views of the National Academy of Sciences. Biographical Memoir COPYRIGHT 1997 NATIONAL ACADEMIES PRESS WASHINGTON D.C. ALLEN NEWELL March 19, 1927–July 19, 1992 BY HERBERT A. SIMON ITH THE DEATH from cancer on July 19, 1992, of Allen WNewell the field of artificial intelligence lost one of its premier scientists, who was at the forefront of the field from its first stirrings to the time of his death and whose research momentum had not shown the slightest diminu- tion up to the premature end of his career. The history of his scientific work is partly my history also, during forty years of friendship and nearly twenty of collaboration, as well as the history of the late J. C. (Cliff) Shaw, a longtime colleague; but I will strive to make this account Allen-cen- tric and not intrude myself too far into it. I hope I will be pardoned if I occasionally fail.1 If you asked Allen Newell what he was, he would say, “I am a scientist.” He played that role almost every waking hour of every day of his adult life. How would he have answered the question, “What kind of scientist?” We hu- mans have long been obsessed with four great questions: the nature of matter, the origins of the universe, the nature of life, the workings of mind. -
Computer Science and the Organization of White-Collar Work, 1945-1975
Post-Industrial Engineering: Computer Science and the Organization of White-Collar Work, 1945-1975 by Andrew Benedict Mamo A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in History in the Graduate Division of the University of California, Berkeley Committee in charge: Professor Cathryn Carson, Chair Professor David Hollinger Professor David Winickoff Spring 2011 Post-Industrial Engineering: Computer Science and the Organization of White-Collar Work, 1945-1975 © 2011 by Andrew Benedict Mamo Abstract Post-Industrial Engineering: Computer Science and the Organization of White-Collar Work, 1945-1975 by Andrew Benedict Mamo Doctor of Philosophy in History University of California, Berkeley Professor Cathryn Carson, Chair The development of computing after the Second World War involved a fundamental reassessment of information, communication, knowledge — and work. No merely technical project, it was prompted in part by the challenges of industrial automation and the shift toward white-collar work in mid-century America. This dissertation therefore seeks out the connections between technical research projects and organization-theory analyses of industrial management in the Cold War years. Rather than positing either a model of technological determinism or one of social construction, it gives a more nuanced description by treating the dynamics as one of constant social and technological co-evolution. This dissertation charts the historical development of what it has meant to work with computers by examining the deep connections between technologists and mid-century organization theorists from the height of managerialism in the 1940s through the decline of the “liberal consensus” in the 1970s. Computing was enmeshed in ongoing debates concerning automation and the relationship between human labor and that of machines. -
A Revisionist History of Connectionism
History: The Past Seite 1 von 5 Copyright 1997, All rights reserved. A Revisionist History of Connectionism Istvan S. N. Berkeley Bibliography According to the standard (recent) history of connectionism (see for example the accounts offered by Hecht-Nielsen (1990: pp. 14-19) and Dreyfus and Dreyfus (1988), or Papert's (1988: pp. 3-4) somewhat whimsical description), in the early days of Classical Computational Theory of Mind (CCTM) based AI research, there was also another allegedly distinct approach, one based upon network models. The work on network models seems to fall broadly within the scope of the term 'connectionist' (see Aizawa 1992), although the term had yet to be coined at the time. These two approaches were "two daughter sciences" according to Papert (1988: p. 3). The fundamental difference between these two 'daughters', lay (according to Dreyfus and Dreyfus (1988: p. 16)) in what they took to be the paradigm of intelligence. Whereas the early connectionists took learning to be fundamental, the traditional school concentrated upon problem solving. Although research on network models initially flourished along side research inspired by the CCTM, network research fell into a rapid decline in the late 1960's. Minsky (aided and abetted by Papert) is often credited with having personally precipitated the demise of research in network models, which marked the end of the first phase of connectionist research. Hecht-Nielson (1990: pp. 16-17) describes the situation (as it is presented in standard versions of the early history of connectionism)