Connectionism: Past, Present, and Future Jordan B
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Dynamic Behavior of Constained Back-Propagation Networks
642 Chauvin Dynamic Behavior of Constrained Back-Propagation Networks Yves Chauvin! Thomson-CSF, Inc. 630 Hansen Way, Suite 250 Palo Alto, CA. 94304 ABSTRACT The learning dynamics of the back-propagation algorithm are in vestigated when complexity constraints are added to the standard Least Mean Square (LMS) cost function. It is shown that loss of generalization performance due to overtraining can be avoided when using such complexity constraints. Furthermore, "energy," hidden representations and weight distributions are observed and compared during learning. An attempt is made at explaining the results in terms of linear and non-linear effects in relation to the gradient descent learning algorithm. 1 INTRODUCTION It is generally admitted that generalization performance of back-propagation net works (Rumelhart, Hinton & Williams, 1986) will depend on the relative size ofthe training data and of the trained network. By analogy to curve-fitting and for theo retical considerations, the generalization performance of the network should de crease as the size of the network and the associated number of degrees of freedom increase (Rumelhart, 1987; Denker et al., 1987; Hanson & Pratt, 1989). This paper examines the dynamics of the standard back-propagation algorithm (BP) and of a constrained back-propagation variation (CBP), designed to adapt the size of the network to the training data base. The performance, learning dynamics and the representations resulting from the two algorithms are compared. 1. Also in the Psychology Department, Stanford University, Stanford, CA. 94305 Dynamic Behavior of Constrained Back-Propagation Networks 643 2 GENERALIZATION PERFORM:ANCE 2.1 STANDARD BACK-PROPAGATION In Chauvin (In Press). the generalization performance of a back-propagation net work was observed for a classification task from spectrograms into phonemic cate gories (single speaker. -
ARCHITECTS of INTELLIGENCE for Xiaoxiao, Elaine, Colin, and Tristan ARCHITECTS of INTELLIGENCE
MARTIN FORD ARCHITECTS OF INTELLIGENCE For Xiaoxiao, Elaine, Colin, and Tristan ARCHITECTS OF INTELLIGENCE THE TRUTH ABOUT AI FROM THE PEOPLE BUILDING IT MARTIN FORD ARCHITECTS OF INTELLIGENCE Copyright © 2018 Packt Publishing All rights reserved. No part of this book may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, without the prior written permission of the publisher, except in the case of brief quotations embedded in critical articles or reviews. Every effort has been made in the preparation of this book to ensure the accuracy of the information presented. However, the information contained in this book is sold without warranty, either express or implied. Neither the author, nor Packt Publishing or its dealers and distributors, will be held liable for any damages caused or alleged to have been caused directly or indirectly by this book. Packt Publishing has endeavored to provide trademark information about all of the companies and products mentioned in this book by the appropriate use of capitals. However, Packt Publishing cannot guarantee the accuracy of this information. Acquisition Editors: Ben Renow-Clarke Project Editor: Radhika Atitkar Content Development Editor: Alex Sorrentino Proofreader: Safis Editing Presentation Designer: Sandip Tadge Cover Designer: Clare Bowyer Production Editor: Amit Ramadas Marketing Manager: Rajveer Samra Editorial Director: Dominic Shakeshaft First published: November 2018 Production reference: 2201118 Published by Packt Publishing Ltd. Livery Place 35 Livery Street Birmingham B3 2PB, UK ISBN 978-1-78913-151-2 www.packt.com Contents Introduction ........................................................................ 1 A Brief Introduction to the Vocabulary of Artificial Intelligence .......10 How AI Systems Learn ........................................................11 Yoshua Bengio .....................................................................17 Stuart J. -
Cogsci Little Red Book.Indd
OUTSTANDING QUESTIONS IN COGNITIVE SCIENCE A SYMPOSIUM HONORING TEN YEARS OF THE DAVID E. RUMELHART PRIZE IN COGNITIVE SCIENCE August 14, 2010 Portland Oregon, USA OUTSTANDING QUESTIONS IN COGNITIVE SCIENCE A SYMPOSIUM HONORING TEN YEARS OF THE DAVID E. RUMELHART PRIZE IN COGNITIVE SCIENCE August 14, 2010 Portland Oregon, USA David E. Rumelhart David E. Rumelhart ABOUT THE RUMELHART PRIZE AND THE 10 YEAR SYMPOSIUM At the August 2000 meeting of the Cognitive Science Society, Drs. James L. McClelland and Robert J. Glushko presented the initial plan to honor the intellectual contributions of David E. Rumelhart to Cognitive Science. Rumelhart had retired from Stanford University in 1998, suffering from Pick’s disease, a degenerative neurological illness. The plan involved honoring Rumelhart through an annual prize of $100,000, funded by the Robert J. Glushko and Pamela Samuelson Foundation. McClelland was a close collaborator of Rumelhart, and, together, they had written numerous articles and books on parallel distributed processing. Glushko, who had been Rumelhart’s Ph.D student in the late 1970s and a Silicon Valley entrepreneur in the 1990s, is currently an adjunct professor at the University of California – Berkeley. The David E. Rumelhart prize was conceived to honor outstanding research in formal approaches to human cognition. Rumelhart’s own seminal contributions to cognitive science included both connectionist and symbolic models, employing both computational and mathematical tools. These contributions progressed from his early work on analogies and story grammars to the development of backpropagation and the use of parallel distributed processing to model various cognitive abilities. Critically, Rumelhart believed that future progress in cognitive science would depend upon researchers being able to develop rigorous, formal theories of mental structures and processes. -
Warren Mcculloch and the British Cyberneticians
Warren McCulloch and the British cyberneticians Article (Accepted Version) Husbands, Phil and Holland, Owen (2012) Warren McCulloch and the British cyberneticians. Interdisciplinary Science Reviews, 37 (3). pp. 237-253. ISSN 0308-0188 This version is available from Sussex Research Online: http://sro.sussex.ac.uk/id/eprint/43089/ This document is made available in accordance with publisher policies and may differ from the published version or from the version of record. If you wish to cite this item you are advised to consult the publisher’s version. Please see the URL above for details on accessing the published version. Copyright and reuse: Sussex Research Online is a digital repository of the research output of the University. Copyright and all moral rights to the version of the paper presented here belong to the individual author(s) and/or other copyright owners. To the extent reasonable and practicable, the material made available in SRO has been checked for eligibility before being made available. Copies of full text items generally can be reproduced, displayed or performed and given to third parties in any format or medium for personal research or study, educational, or not-for-profit purposes without prior permission or charge, provided that the authors, title and full bibliographic details are credited, a hyperlink and/or URL is given for the original metadata page and the content is not changed in any way. http://sro.sussex.ac.uk Warren McCulloch and the British Cyberneticians1 Phil Husbands and Owen Holland Dept. Informatics, University of Sussex Abstract Warren McCulloch was a significant influence on a number of British cyberneticians, as some British pioneers in this area were on him. -
Kybernetik in Urbana
Kybernetik in Urbana Ein Gespräch zwischen Paul Weston, Jan Müggenburg und James Andrew Hutchinson1 Paul Ernest Weston was born on February 3rd 1935 in Garland, Maine. After stu- dying physics at Wesleyan University in Middletown (Connecticut), 1952 through 1956, Paul Weston came to the University of Illinois in summer 1956, receiving his M.S. in August 1958. The following autumn, as a graduate student of physics, he attended a seminar led by Heinz von Foerster under the promising title „Cyberne- tics“.2 In June 1959 Paul joins the newly founded Biological Computer Laboratory (BCL) as a half time research assistant employed by the Electrical Engineering Department. Through roughly 1962, Weston’s primary emphasis was upon simulation of the processes of perception, with work in neuron models and pattern recognition. Together with Murray Babcock, Weston was one of the most prominent creators of prototypes built at the BCL. Already in 1961 Weston becomes widely known as the inventor of the NumaRete, a parallel operated pattern recognition machine, which played an important role in the public perception of the BCL through out the fol- lowing years.3 After 1964 Weston continues his research at BCL on the basis of a full position as a research assistant and starts to focus on information processing, particularly inte- rested in problems associated with natural-language interaction with machine. As a member of the Committee for Cognitive Studies Weston is also increasingly interes- ted in the potentials of how the new information technologies could shape society. In June 1970 Weston receives his PhD for developing a new and more efficient data structure called CYLINDER.4 After the BCL was closed in 1975 Weston continued to work as a Research Associate at the Department of Electrical Engineering and the Coordinated Science Laboratory of the University of Illinois. -
1011 Neurons } 1014 Synapses the Cybernetics Group
History 329/SI 311/RCSSCI 360 Computers and the Internet: A global history Week 6: Computing and Cybernetics in the Soviet Union Today } Review } Cybernetics: minds, brains, and machines } Soviet computing and cybernetics } Soviet economic planning } Next time Review: SAGE } Origin: Whirlwind digital computer project, MIT } SAGE = Semi-Automatic Ground Environment } Computer-controlled air defense of lower 48 states } Networked over telephone lines } Duplexed (two computers for high reliability) } Tube-based central processors made by IBM } Magnetic core memory } First truly large software development } Served as a pattern for many subsequent military projects } Major factor in US and IBM dominance of commercial computer markets by late1950s Cybernetics: minds, brains, and machines Key figures } Norbert Wiener } cybernetics } Warren McCulloch & Walter Pitts } neural nets } John von Neumann } brain modeling, cellular automata, biological metaphors } Frank Rosenblatt } perceptrons The Cybernetics Group } Norbert Wiener, MIT } WWII anti-aircraft problems } Servo/organism analogies } “Behavior, Purpose, and Teleology” (1943) } Information as measurable quantity } Feedback: circular self-corrective cycles BEHAVIOR, PURPOSE AND TELEOLOGY 21 Examples of predictions of higher order are shooting with a sling or with a bow and arrow. Predictive behavior requires the discrimination of at least two coordinates, a temporal and at least one spatial axis. Prediction will be more effective and flexible, however, if the behaving object can respond to changes in more than one spatial coordinate. The sensory receptors of an organism, or the corresponding elements of a machine, may therefore limit the predictive behavior. Thus, a bloodhound follows a trail, that is, it does not show any predictive behavior in trailing, because a chemical, olfactory input reports only spatial information: distance, as indicated by intensity. -
Journal of Sociocybernetics
JOURNAL OF SOCIOCYBERNETICS _________________________________________________________________ Volume 3 Number 1 Spring/Summer 2002 Official Journal of the Research Committee on Sociocybernetics (RC51) of the International Sociological Association JOURNAL OF SOCIOCYBERNETICS www.unizar.es/sociocybernetics/ Editor Richard E. Lee Newsletter Cor Van Dijkum Felix Geyer Editorial Board Mike Byron Tessaleno Devezas Jorge González Bernd R. Hornung Chaime Marcuello Vessela Misheva Philip Nikolopoulos Bernard Scott Mike Terpstra ____________________________________________________________________________ The JOURNAL OF SOCIOCYBERNETICS (ISSN 1607-8667) is an electronic journal published biannually--Spring/Summer and Fall/Winter--by the Research Committee on Sociocybernetics of the International Sociological Association. MANUSCRIPT submissions should be sent electronically (in MSWord or Rich Text File format) to each of the editors: Richard E. Lee [email protected], Felix Geyer, [email protected], and Cor van Dijkum, [email protected]. In general, please follow the Chicago Manuel of Style; citations and bibliography should follow the current journal style (APA). Normally, articles should be original texts of no more than 6000 words, although longer articles will be considered in exceptional circumstances. The Journal looks for submissions that are innovative and apply principles of General Systems Theory and Cybernetics to the social sciences, broadly conceived. COPYRIGHT remains the property of authors. Permission to reprint must be obtained from the authors and the contents of JoS cannot be copied for commercial purposes. JoS does, however, reserve the right to future reproduction of articles in hard copy, portable document format (.pdf), or HTML editions of JoS. iii SOCIOCYBERNETICS traces its intellectual roots to the rise of a panoply of new approaches to scientific inquiry beginning in the 1940's. -
Introspection:Accelerating Neural Network Training by Learning Weight Evolution
Published as a conference paper at ICLR 2017 INTROSPECTION:ACCELERATING NEURAL NETWORK TRAINING BY LEARNING WEIGHT EVOLUTION Abhishek Sinha∗ Mausoom Sarkar Department of Electronics and Electrical Comm. Engg. Adobe Systems Inc, Noida IIT Kharagpur Uttar Pradesh,India West Bengal, India msarkar at adobe dot com abhishek.sinha94 at gmail dot com Aahitagni Mukherjee∗ Balaji Krishnamurthy Department of Computer Science Adobe Systems Inc, Noida IIT Kanpur Uttar Pradesh,India Uttar Pradesh, India kbalaji at adobe dot com ahitagnimukherjeeam at gmail dot com ABSTRACT Neural Networks are function approximators that have achieved state-of-the-art accuracy in numerous machine learning tasks. In spite of their great success in terms of accuracy, their large training time makes it difficult to use them for various tasks. In this paper, we explore the idea of learning weight evolution pattern from a simple network for accelerating training of novel neural networks. We use a neural network to learn the training pattern from MNIST classifi- cation and utilize it to accelerate training of neural networks used for CIFAR-10 and ImageNet classification. Our method has a low memory footprint and is computationally efficient. This method can also be used with other optimizers to give faster convergence. The results indicate a general trend in the weight evolution during training of neural networks. 1 INTRODUCTION Deep neural networks have been very successful in modeling high-level abstractions in data. How- ever, training a deep neural network for any AI task is a time-consuming process. This is because a arXiv:1704.04959v1 [cs.LG] 17 Apr 2017 large number of parameters need to be learnt using training examples. -
Constructing Artificial Intelligence Paul Edwards, the Closed World
* c 4 1 v. N > COMPUTERS . discourse »"• "u m com *»» *l't"'tA PAUL N. EDWARDS The Closed World Inside Technology edited by Wiebe E. Bijker, W. Bernard Carlson, and Trevor Pinch Wiebe E. Bijker, Of Bicycles, Bahelites, and Bulbs: Toward a Theory of Sociotechnical Change Wiebe E. Bijker and John Law, editors, Shaping Technology/Building Society: Studies m Sociotechnical Change Stuart S. Blume, Insight and Industry: On the Dynamics of Technological Change in Medicine Louis L. Bucciarelli, Designing Engineers Geoffrey C. Bowker, Science on the Run: Information Management and Industrial Geophysics at Schlumberger, 1920-1940 H. M. Collins, Artificial Experts: Social Knowledge and Intelligent Machines Paul N. Edwards, The Closed World: Computers and the Politics of Discourse in Cold War America Pamela E. Mack, Viewing the Earth: The Social Construction of the Landsat Satellite System Donald MacKenzie, Inventing Accuracy: A Historical Sociology of Nuclear Missile Guidance Donald MacKenzie, Knowing Machines: Essays on Technical Change The Closed World Computers and the Politics of Discourse in Cold War America Paul N. Edwards The MIT Press Cambridge, Massachusetts London, England ©1996 Massachusetts Institute of Technology All rights reserved. No part of this book may be reproduced in any form or by any electronic or mechanical means (including photocopying, recording, or information storage and retrieval) without permission in writing from the publisher. This book was set in Baskerville by Pine Tree Composition, Inc. and printed and bound in the United States of America. Library of Congress Cataloging-in-Publication Data Edwards, Paul N. The closed world : computers and the politics of discourse in Cold War America / Paul N. -
A Defense of Pure Connectionism
City University of New York (CUNY) CUNY Academic Works All Dissertations, Theses, and Capstone Projects Dissertations, Theses, and Capstone Projects 2-2019 A Defense of Pure Connectionism Alex B. Kiefer The Graduate Center, City University of New York How does access to this work benefit ou?y Let us know! More information about this work at: https://academicworks.cuny.edu/gc_etds/3036 Discover additional works at: https://academicworks.cuny.edu This work is made publicly available by the City University of New York (CUNY). Contact: [email protected] A DEFENSE OF PURE CONNECTIONISM by ALEXANDER B. KIEFER A dissertation submitted to the Graduate Faculty in Philosophy in partial fulfillment of the requirements for the degree of Doctor of Philosophy, The City University of New York 2019 © 2019 ALEXANDER B. KIEFER All Rights Reserved ii A Defense of Pure Connectionism By Alexander B. Kiefer This manuscript has been read and accepted by the Graduate Faculty in Philosophy in satisfaction of the dissertation requirement for the degree of Doctor of Philosophy. _________________________ _________________________________________ 1/22/2019 Eric Mandelbaum Chair of Examining Committee _________________________ _________________________________________ 1/22/2019 Nickolas Pappas Executive Officer Supervisory Committee: David Rosenthal Jakob Hohwy Eric Mandelbaum THE CITY UNIVERSITY OF NEW YORK iii ABSTRACT A Defense of Pure Connectionism by Alexander B. Kiefer Advisor: David M. Rosenthal Connectionism is an approach to neural-networks-based cognitive modeling that encompasses the recent deep learning movement in artificial intelligence. It came of age in the 1980s, with its roots in cybernetics and earlier attempts to model the brain as a system of simple parallel processors. -
Generative Linguistics and Neural Networks at 60: Foundation, Friction, and Fusion*
Generative linguistics and neural networks at 60: foundation, friction, and fusion* Joe Pater, University of Massachusetts Amherst October 3, 2018. Abstract. The birthdate of both generative linguistics and neural networks can be taken as 1957, the year of the publication of foundational work by both Noam Chomsky and Frank Rosenblatt. This paper traces the development of these two approaches to cognitive science, from their largely autonomous early development in their first thirty years, through their collision in the 1980s around the past tense debate (Rumelhart and McClelland 1986, Pinker and Prince 1988), and their integration in much subsequent work up to the present. Although this integration has produced a considerable body of results, the continued general gulf between these two lines of research is likely impeding progress in both: on learning in generative linguistics, and on the representation of language in neural modeling. The paper concludes with a brief argument that generative linguistics is unlikely to fulfill its promise of accounting for language learning if it continues to maintain its distance from neural and statistical approaches to learning. 1. Introduction At the beginning of 1957, two men nearing their 29th birthdays published work that laid the foundation for two radically different approaches to cognitive science. One of these men, Noam Chomsky, continues to contribute sixty years later to the field that he founded, generative linguistics. The book he published in 1957, Syntactic Structures, has been ranked as the most influential work in cognitive science from the 20th century.1 The other one, Frank Rosenblatt, had by the late 1960s largely moved on from his research on perceptrons – now called neural networks – and died tragically young in 1971. -
A Logical Calculus of the Ideas Immanent in Nervous Activity*
Bulletin of Mothemnticnl Biology Vol. 52, No. l/2. pp. 99-115. 1990. oo92-824OjW$3.OO+O.MI Printed in Great Britain. Pergamon Press plc Society for Mathematical Biology A LOGICAL CALCULUS OF THE IDEAS IMMANENT IN NERVOUS ACTIVITY* n WARREN S. MCCULLOCH AND WALTER PITTS University of Illinois, College of Medicine, Department of Psychiatry at the Illinois Neuropsychiatric Institute, University of Chicago, Chicago, U.S.A. Because of the “all-or-none” character of nervous activity, neural events and the relations among them can be treated by means of propositional logic. It is found that the behavior of every net can be described in these terms, with the addition of more complicated logical means for nets containing circles; and that for any logical expression satisfying certain conditions, one can find a net behaving in the fashion it describes. It is shown that many particular choices among possible neurophysiological assumptions are equivalent, in the sense that for every net behaving under one assumption, there exists another net which behaves under the other and gives the same results, although perhaps not in the same time. Various applications of the calculus are discussed. 1. Introduction. Theoretical neurophysiology rests on certain cardinal assumptions. The nervous system is a net of neurons, each having a soma and an axon. Their adjunctions, or synapses, are always between the axon of one neuron and the soma of another. At any instant a neuron has some threshold, which excitation must exceed to initiate an impulse. This, except for the fact and the time of its occurence, is determined by the neuron, not by the excitation.