GPT-3: Its Nature, Scope, Limits, and Consequences
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Can Machines Think?
Sandringham School Faculty of Computer Science Can Machines Think? This is a question posed by famous English Computer Scientist Alan Turing in Intelligent machines and programmes are judged. 1950. His ‘Turing Test’ has become the benchmark by which Artificially Recently, the Turing Test hit the news when a programme that simulated a 13 year old Ukrainian boy (called Eugene Goostman) managed to convince a third of testers that they were talking to a real human. http://www.bbc.co.uk/news/technology-27762088 http://www.reading.ac.uk/news-and-events/releases/PR583836.aspx http://www.alicebot.org/chatbots3/Eugene.pdf Your summer assignment is as follows: 1. 2. DesigningInvestigate a who fully Alan functional Turing Chatbot was and that explain can process in detail human what the language ‘Turing is oneTest’ of is the and most how difficult it is important tasks in in Computer the world Science. of Artificial Why Intelligence is this, what are the challenges? How did the Eugene Goostman bot overcome these? 3. Time to write your own Chatbot. In any language you choose (e.g. Python, Scratch, Java etc.), write a simple Chatbot. Clearly you will have to limit the questions that your bot can ask and the responses that it will Chatbot called Eliza at : https://groklearning.com/csedweek/hoc-eliza/ 4. What‘understand’! are the positiveIf you are implications totally stuck, for see the getting development started ofwith Artificial a simple Intelligence? What are the dangers to society of truly intelligent machines? Your assignment should be at least 4 pages of A4, excluding the programme listing. -
Generative Adversarial Networks (Gans)
Generative Adversarial Networks (GANs) The coolest idea in Machine Learning in the last twenty years - Yann Lecun Generative Adversarial Networks (GANs) 1 / 39 Overview Generative Adversarial Networks (GANs) 3D GANs Domain Adaptation Generative Adversarial Networks (GANs) 2 / 39 Introduction Generative Adversarial Networks (GANs) 3 / 39 Supervised Learning Find deterministic function f: y = f(x), x:data, y:label Generative Adversarial Networks (GANs) 4 / 39 Unsupervised Learning "Most of human and animal learning is unsupervised learning. If intelligence was a cake, unsupervised learning would be the cake, supervised learning would be the icing on the cake, and reinforcement learning would be the cherry on the cake. We know how to make the icing and the cherry, but we do not know how to make the cake. We need to solve the unsupervised learning problem before we can even think of getting to true AI." - Yann Lecun "You cannot predict what you cannot understand" - Anonymous Generative Adversarial Networks (GANs) 5 / 39 Unsupervised Learning More challenging than supervised learning. No label or curriculum. Some NN solutions: Boltzmann machine AutoEncoder Generative Adversarial Networks Generative Adversarial Networks (GANs) 6 / 39 Unsupervised Learning vs Generative Model z = f(x) vs. x = g(z) P(z|x) vs P(x|z) Generative Adversarial Networks (GANs) 7 / 39 Autoencoders Stacked Autoencoders Use data itself as label Generative Adversarial Networks (GANs) 8 / 39 Autoencoders Denosing Autoencoders Generative Adversarial Networks (GANs) 9 / 39 Variational Autoencoder Generative Adversarial Networks (GANs) 10 / 39 Variational Autoencoder Results Generative Adversarial Networks (GANs) 11 / 39 Generative Adversarial Networks Ian Goodfellow et al, "Generative Adversarial Networks", 2014. -
Language Technologies Past Present, and Future
Language technologies past present, and future Christopher Potts CSLI Summer Internship Program July 21, 2017 Many slides joint work with Bill MacCartney: http://web.stanford.edu/class/cs224u/ Hype and hand-wringing • Jürgen Schmidhuber: “We are on the verge not of another industrial revolution, but a new form of life, more like the big bang.” [link] • Elon Musk: “AI is a fundamental existential risk for human civilization, and I don't think people fully appreciate that.” [link] • October 2016: “Microsoft has made a major breakthrough in speech recognition, creating a technology that recognizes the words in a conversation as well as a person does.” [link] Two perspectives Overview • What is understanding? • A brief history of language technologies • Language technologies of the past • Language technologies of today • Current approaches and prospects • Predictions about the future! Readings and other background • Percy Liang: Talking to computers in natural language • Levesque: On our best behaviour • Mitchell: Reading the Web: A breakthrough goal for AI • Podcast: The challenge and promise of artificial intelligence • Podcast: Hal Daume on Talking Machines • Stanford CS224u, Natural language understanding What is understanding? To understand a statement is to: • determine its truth (with justification) • calculate its entailments • take appropriate action in light of it • translate it into another language • … The Turing Test (Turing 1950) Turing replaced “Can machines think?”, which he regarded as “too meaningless to deserve discussion” (p. 442), with the question whether an interrogator could be tricked into thinking that a machine was a human using only conversation (no visuals, no demands for physical performance, etc.). Some of the objections Turing anticipates • “Thinking is a function of man’s immortal soul. -
Oliver Knill: March 2000 Literature: Peter Norvig, Paradigns of Artificial Intelligence Programming Daniel Juravsky and James Martin, Speech and Language Processing
ENTRY ARTIFICIAL INTELLIGENCE [ENTRY ARTIFICIAL INTELLIGENCE] Authors: Oliver Knill: March 2000 Literature: Peter Norvig, Paradigns of Artificial Intelligence Programming Daniel Juravsky and James Martin, Speech and Language Processing Adaptive Simulated Annealing [Adaptive Simulated Annealing] A language interface to a neural net simulator. artificial intelligence [artificial intelligence] (AI) is a field of computer science concerned with the concepts and methods of symbolic knowledge representation. AI attempts to model aspects of human thought on computers. Aspectrs of AI: computer vision • language processing • pattern recognition • expert systems • problem solving • roboting • optical character recognition • artificial life • grammars • game theory • Babelfish [Babelfish] Online translation system from Systran. Chomsky [Chomsky] Noam Chomsky is a pioneer in formal language theory. He is MIT Professor of Linguistics, Linguistic Theory, Syntax, Semantics and Philosophy of Language. Eliza [Eliza] One of the first programs to feature English output as well as input. It was developed by Joseph Weizenbaum at MIT. The paper appears in the January 1966 issue of the "Communications of the Association of Computing Machinery". Google [Google] A search engine emerging at the end of the 20'th century. It has AI features, allows not only to answer questions by pointing to relevant webpages but can also do simple tasks like doing arithmetic computations, convert units, read the news or find pictures with some content. GPS [GPS] General Problem Solver. A program developed in 1957 by Alan Newell and Herbert Simon. The aim was to write a single computer program which could solve any problem. One reason why GPS was destined to fail is now at the core of computer science. -
The Points of Viewing Theory for Engaged Learning with Digital Media
The Points of Viewing Theory for Engaged Learning with Digital Media Ricki Goldman New York University John Black Columbia University John W. Maxwell Simon Fraser University Jan J. L. Plass New York University Mark J. Keitges University of Illinois, Urbana Champaign - 1 - Introduction Theories are dangerous things. All the same we must risk making one this afternoon since we are going to discuss modern tendencies. Directly we speak of tendencies or movements we commit to, the belief that there is some force, influence, outer pressure that is strong enough to stamp itself upon a whole group of different writers so that all their writing has a certain common likeness. — Virginia Woolff, The Leaning Tower, lecture delivered to the Workers' Educational Association, Brighton (May 1940). With full acknowledgement of the warning from the 1940 lecture by Virginia Woolf, this chapter begins by presenting a theory of mind, knowing only too well, that “a whole group of different” learning theorists cannot find adequate coverage under one umbrella. Nor should they. However, there is a movement occurring, a form of social activism created by the affordances of social media, an infrastructure that was built incrementally during two to three decades of hard scholarly research that brought us to this historic time and place. To honor the convergence of theories and technologies, this paper proposes the Points of Viewing Theory to provide researchers, teachers, and the public with an opportunity to discuss and perhaps change the epistemology of education from its formal structures to more do-it-yourself learning environments that dig deeper and better into content knowledge. -
Much Has Been Written About the Turing Test in the Last Few Years, Some of It
1 Much has been written about the Turing Test in the last few years, some of it preposterously off the mark. People typically mis-imagine the test by orders of magnitude. This essay is an antidote, a prosthesis for the imagination, showing how huge the task posed by the Turing Test is, and hence how unlikely it is that any computer will ever pass it. It does not go far enough in the imagination-enhancement department, however, and I have updated the essay with a new postscript. Can Machines Think?1 Can machines think? This has been a conundrum for philosophers for years, but in their fascination with the pure conceptual issues they have for the most part overlooked the real social importance of the answer. It is of more than academic importance that we learn to think clearly about the actual cognitive powers of computers, for they are now being introduced into a variety of sensitive social roles, where their powers will be put to the ultimate test: In a wide variety of areas, we are on the verge of making ourselves dependent upon their cognitive powers. The cost of overestimating them could be enormous. One of the principal inventors of the computer was the great 1 Originally appeared in Shafto, M., ed., How We Know (San Francisco: Harper & Row, 1985). 2 British mathematician Alan Turing. It was he who first figured out, in highly abstract terms, how to design a programmable computing device--what we not call a universal Turing machine. All programmable computers in use today are in essence Turing machines. -
Passing the Turing Test Does Not Mean the End of Humanity
Cogn Comput (2016) 8:409–419 DOI 10.1007/s12559-015-9372-6 Passing the Turing Test Does Not Mean the End of Humanity 1 1 Kevin Warwick • Huma Shah Received: 18 September 2015 / Accepted: 20 November 2015 / Published online: 28 December 2015 Ó The Author(s) 2015. This article is published with open access at Springerlink.com Abstract In this paper we look at the phenomenon that is we do wish to dispel, however, is the assumption which the Turing test. We consider how Turing originally intro- links passing the Turing test with the achievement for duced his imitation game and discuss what this means in a machines of human-like or human-level intelligence. practical scenario. Due to its popular appeal we also look Unfortunately the assumed chain of events which means into different representations of the test as indicated by that passing the Turing test sounds the death knell for numerous reviewers. The main emphasis here, however, is humanity appears to have become engrained in the thinking to consider what it actually means for a machine to pass the in certain quarters. One interesting corollary of this is that Turing test and what importance this has, if any. In par- when it was announced in 2014 that the Turing test had ticular does it mean that, as Turing put it, a machine can been finally passed [39] there was an understandable ‘‘think’’. Specifically we consider claims that passing the response from those same quarters that it was not possible Turing test means that machines will have achieved for such an event to have occurred, presumably because we human-like intelligence and as a consequence the singu- were still here in sterling health to both make and debate larity will be upon us in the blink of an eye. -
Turing Test Does Not Work in Theory but in Practice
Int'l Conf. Artificial Intelligence | ICAI'15 | 433 Turing test does not work in theory but in practice Pertti Saariluoma1 and Matthias Rauterberg2 1Information Technology, University of Jyväskylä, Jyväskylä, Finland 2Industrial Design, Eindhoven University of Technology, Eindhoven, The Netherlands Abstract - The Turing test is considered one of the most im- Essentially, the Turing test is an imitation game. Like any portant thought experiments in the history of AI. It is argued good experiment, it has two conditions. In the case of control that the test shows how people think like computers, but is this conditions, it is assumed that there is an opaque screen. On actually true? In this paper, we discuss an entirely new per- one side of the screen is an interrogator whose task is to ask spective. Scientific languages have their foundational limita- questions and assess the nature of the answers. On the other tions, for example, in their power of expression. It is thus pos- side, there are two people, A and B. The task of A and B is to sible to discuss the limitations of formal concepts and theory answer the questions, and the task of the interrogator is to languages. In order to represent real world phenomena in guess who has given the answer. In the case of experimental formal concepts, formal symbols must be given semantics and conditions, B is replaced by a machine (computer) and again, information contents; that is, they must be given an interpreta- the interrogator must decide whether it was the human or the tion. They key argument is that it is not possible to express machine who answered the questions. -
THE TURING TEST RELIES on a MISTAKE ABOUT the BRAIN For
THE TURING TEST RELIES ON A MISTAKE ABOUT THE BRAIN K. L. KIRKPATRICK Abstract. There has been a long controversy about how to define and study intel- ligence in machines and whether machine intelligence is possible. In fact, both the Turing Test and the most important objection to it (called variously the Shannon- McCarthy, Blockhead, and Chinese Room arguments) are based on a mistake that Turing made about the brain in his 1948 paper \Intelligent Machinery," a paper that he never published but whose main assumption got embedded in his famous 1950 paper and the celebrated Imitation Game. In this paper I will show how the mistake is a false dichotomy and how it should be fixed, to provide a solid foundation for a new understanding of the brain and a new approach to artificial intelligence. In the process, I make an analogy between the brain and the ribosome, machines that translate information into action, and through this analogy I demonstrate how it is possible to go beyond the purely information-processing paradigm of computing and to derive meaning from syntax. For decades, the metaphor of the brain as an information processor has dominated both neuroscience and artificial intelligence research and allowed the fruitful appli- cations of Turing's computation theory and Shannon's information theory in both fields. But this metaphor may be leading us astray, because of its limitations and the deficiencies in our understanding of the brain and AI. In this paper I will present a new metaphor to consider, of the brain as both an information processor and a producer of physical effects. -
Dynamic Factor Graphs for Time Series Modeling
Dynamic Factor Graphs for Time Series Modeling Piotr Mirowski and Yann LeCun Courant Institute of Mathematical Sciences, New York University, 719 Broadway, New York, NY 10003 USA {mirowski,yann}@cs.nyu.edu http://cs.nyu.edu/∼mirowski/ Abstract. This article presents a method for training Dynamic Fac- tor Graphs (DFG) with continuous latent state variables. A DFG in- cludes factors modeling joint probabilities between hidden and observed variables, and factors modeling dynamical constraints on hidden vari- ables. The DFG assigns a scalar energy to each configuration of hidden and observed variables. A gradient-based inference procedure finds the minimum-energy state sequence for a given observation sequence. Be- cause the factors are designed to ensure a constant partition function, they can be trained by minimizing the expected energy over training sequences with respect to the factors’ parameters. These alternated in- ference and parameter updates can be seen as a deterministic EM-like procedure. Using smoothing regularizers, DFGs are shown to reconstruct chaotic attractors and to separate a mixture of independent oscillatory sources perfectly. DFGs outperform the best known algorithm on the CATS competition benchmark for time series prediction. DFGs also suc- cessfully reconstruct missing motion capture data. Key words: factor graphs, time series, dynamic Bayesian networks, re- current networks, expectation-maximization 1 Introduction 1.1 Background Time series collected from real-world phenomena are often an incomplete picture of a complex underlying dynamical process with a high-dimensional state that cannot be directly observed. For example, human motion capture data gives the positions of a few markers that are the reflection of a large number of joint angles with complex kinematic and dynamical constraints. -
Artificial Intelligence and Big Data – Innovation Landscape Brief
ARTIFICIAL INTELLIGENCE AND BIG DATA INNOVATION LANDSCAPE BRIEF © IRENA 2019 Unless otherwise stated, material in this publication may be freely used, shared, copied, reproduced, printed and/or stored, provided that appropriate acknowledgement is given of IRENA as the source and copyright holder. Material in this publication that is attributed to third parties may be subject to separate terms of use and restrictions, and appropriate permissions from these third parties may need to be secured before any use of such material. ISBN 978-92-9260-143-0 Citation: IRENA (2019), Innovation landscape brief: Artificial intelligence and big data, International Renewable Energy Agency, Abu Dhabi. ACKNOWLEDGEMENTS This report was prepared by the Innovation team at IRENA’s Innovation and Technology Centre (IITC) with text authored by Sean Ratka, Arina Anisie, Francisco Boshell and Elena Ocenic. This report benefited from the input and review of experts: Marc Peters (IBM), Neil Hughes (EPRI), Stephen Marland (National Grid), Stephen Woodhouse (Pöyry), Luiz Barroso (PSR) and Dongxia Zhang (SGCC), along with Emanuele Taibi, Nadeem Goussous, Javier Sesma and Paul Komor (IRENA). Report available online: www.irena.org/publications For questions or to provide feedback: [email protected] DISCLAIMER This publication and the material herein are provided “as is”. All reasonable precautions have been taken by IRENA to verify the reliability of the material in this publication. However, neither IRENA nor any of its officials, agents, data or other third- party content providers provides a warranty of any kind, either expressed or implied, and they accept no responsibility or liability for any consequence of use of the publication or material herein. -
Energy-Based Generative Adversarial Network
Published as a conference paper at ICLR 2017 ENERGY-BASED GENERATIVE ADVERSARIAL NET- WORKS Junbo Zhao, Michael Mathieu and Yann LeCun Department of Computer Science, New York University Facebook Artificial Intelligence Research fjakezhao, mathieu, [email protected] ABSTRACT We introduce the “Energy-based Generative Adversarial Network” model (EBGAN) which views the discriminator as an energy function that attributes low energies to the regions near the data manifold and higher energies to other regions. Similar to the probabilistic GANs, a generator is seen as being trained to produce contrastive samples with minimal energies, while the discriminator is trained to assign high energies to these generated samples. Viewing the discrimi- nator as an energy function allows to use a wide variety of architectures and loss functionals in addition to the usual binary classifier with logistic output. Among them, we show one instantiation of EBGAN framework as using an auto-encoder architecture, with the energy being the reconstruction error, in place of the dis- criminator. We show that this form of EBGAN exhibits more stable behavior than regular GANs during training. We also show that a single-scale architecture can be trained to generate high-resolution images. 1I NTRODUCTION 1.1E NERGY-BASED MODEL The essence of the energy-based model (LeCun et al., 2006) is to build a function that maps each point of an input space to a single scalar, which is called “energy”. The learning phase is a data- driven process that shapes the energy surface in such a way that the desired configurations get as- signed low energies, while the incorrect ones are given high energies.