Oliver Knill: March 2000 Literature: Peter Norvig, Paradigns of Artificial Intelligence Programming Daniel Juravsky and James Martin, Speech and Language Processing
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
The Points of Viewing Theory for Engaged Learning with Digital Media
The Points of Viewing Theory for Engaged Learning with Digital Media Ricki Goldman New York University John Black Columbia University John W. Maxwell Simon Fraser University Jan J. L. Plass New York University Mark J. Keitges University of Illinois, Urbana Champaign - 1 - Introduction Theories are dangerous things. All the same we must risk making one this afternoon since we are going to discuss modern tendencies. Directly we speak of tendencies or movements we commit to, the belief that there is some force, influence, outer pressure that is strong enough to stamp itself upon a whole group of different writers so that all their writing has a certain common likeness. — Virginia Woolff, The Leaning Tower, lecture delivered to the Workers' Educational Association, Brighton (May 1940). With full acknowledgement of the warning from the 1940 lecture by Virginia Woolf, this chapter begins by presenting a theory of mind, knowing only too well, that “a whole group of different” learning theorists cannot find adequate coverage under one umbrella. Nor should they. However, there is a movement occurring, a form of social activism created by the affordances of social media, an infrastructure that was built incrementally during two to three decades of hard scholarly research that brought us to this historic time and place. To honor the convergence of theories and technologies, this paper proposes the Points of Viewing Theory to provide researchers, teachers, and the public with an opportunity to discuss and perhaps change the epistemology of education from its formal structures to more do-it-yourself learning environments that dig deeper and better into content knowledge. -
UC Berkeley Previously Published Works
UC Berkeley UC Berkeley Previously Published Works Title Building the Second Mind, 1961-1980: From the Ascendancy of ARPA-IPTO to the Advent of Commercial Expert Systems Permalink https://escholarship.org/uc/item/7ck3q4f0 ISBN 978-0-989453-4-6 Author Skinner, Rebecca Elizabeth Publication Date 2013-12-31 eScholarship.org Powered by the California Digital Library University of California Building the Second Mind, 1961-1980: From the Ascendancy of ARPA to the Advent of Commercial Expert Systems copyright 2013 Rebecca E. Skinner ISBN 978 09894543-4-6 Forward Part I. Introduction Preface Chapter 1. Introduction: The Status Quo of AI in 1961 Part II. Twin Bolts of Lightning Chapter 2. The Integrated Circuit Chapter 3. The Advanced Research Projects Agency and the Foundation of the IPTO Chapter 4. Hardware, Systems and Applications in the 1960s Part II. The Belle Epoque of the 1960s Chapter 5. MIT: Work in AI in the Early and Mid-1960s Chapter 6. CMU: From the General Problem Solver to the Physical Symbol System and Production Systems Chapter 7. Stanford University and SRI Part III. The Challenges of 1970 Chapter 8. The Mansfield Amendment, “The Heilmeier Era”, and the Crisis in Research Funding Chapter 9. The AI Culture Wars: the War Inside AI and Academia Chapter 10. The AI Culture Wars: Popular Culture Part IV. Big Ideas and Hardware Improvements in the 1970s invert these and put the hardware chapter first Chapter 11. AI at MIT in the 1970s: The Semantic Fallout of NLR and Vision Chapter 12. Hardware, Software, and Applications in the 1970s Chapter 13. -
An Overview of Artificial Intelligence and Robotics
AN OVERVIEW OF ARTIFICIAL INTELLIGENCE AND ROBOTICS VOLUME1—ARTIFICIAL INTELLIGENCE PART A—THE CORE INGREDIENTS January 1984 Prepared for National Aeronautics and Space Administration Headquarters Washington, D.C. 20546 NATIONAL E'j • OF STA. IDALL L LIBRARY qc ~iy?f NBSIR 83-2687 AN OVERVIEW OF ARTIFICIAL INTELLIGENCE AND ROBOTICS VOLUME1 —ARTIFICIAL INTELLIGENCE PART A—THE CORE INGREDIENTS William B. Gevarter* U.S. DEPARTMENT OF COMMERCE National Bureau of Standards National Engineering Laboratory Center for Manufacturing Engineering Industrial Systems Division Metrology Building, Room A127 Washington, DC 20234 January 1984 Prepared for: National Aeronautics and Space Administration Headquarters Washington, DC 20546 U.S. DEPARTMENT OF COMMERCE, Malcolm Baldrige, Secretary NATIONAL BUREAU OF STANDARDS. Ernest Ambler, Director •Research Associate at the National Bureau of Standards Sponsored by NASA Headquarters PREFACE Artificial Intelligence (AI) is a field with over a quarter of a century of history. However, it wasn’t until the dawn of the decade of the 80’s that AI really burst forth—finding economic and popular acclaim. Intelligent computer programs are now emerging from the laboratory into prac- tical applications. This report endeavors to indicate what AI is, the foundations on which it rests, the techniques utilized, the applications, the participants and finally its state-of-the-art and future trends. Due to the scope of AI, this volume is issued in three parts: Part A: The Core Ingredients I. Artificial Intelligence—What It Is II. The Rise, Fall and Rebirth of AI III. Basic Elements of AI IV. Applications V. The Principal Participants VI. State-of-the-Art VII. -
Inconsistency Robustness for Logic Programs Carl Hewitt
Inconsistency Robustness for Logic Programs Carl Hewitt To cite this version: Carl Hewitt. Inconsistency Robustness for Logic Programs. Inconsistency Robustness, 2015, 978-1- 84890-159. hal-01148496v6 HAL Id: hal-01148496 https://hal.archives-ouvertes.fr/hal-01148496v6 Submitted on 27 Apr 2016 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Copyright Inconsistency Robustness for Logic Programs Carl Hewitt This article is dedicated to Alonzo Church and Stanisław Jaśkowski Abstract This article explores the role of Inconsistency Robustness in the history and theory of Logic Programs. Inconsistency Robustness has been a continually recurring issue in Logic Programs from the beginning including Church's system developed in the early 1930s based on partial functions (defined in the lambda calculus) that he thought would allow development of a general logic without the kind of paradoxes that had plagued earlier efforts by Frege, etc.1 Planner [Hewitt 1969, 1971] was a kind of hybrid between the procedural and logical paradigms in that it featured a procedural embedding of logical sentences in that an implication of the form (p implies q) can be procedurally embedded in the following ways: Forward chaining When asserted p, Assert q When asserted q, Assert p Backward chaining When goal q, SetGoal p When goal p, SetGoal q Developments by different research groups in the fall of 1972 gave rise to a controversy over Logic Programs that persists to this day in the form of following alternatives: 1. -
Social Interaction with Computers: a Sociotechnical Interpretation Of
Social Interaction with Computers: an Interpretation of Weizenbaum’s ELIZA and Her Heritage. Hans Pruijt Erasmus University Rotterdam Submitted for publication in Social Science Computer Review (Pruijt, H., 2006, “Social Interaction with Computers: An Interpretation of Weizenbaum’s ELIZA and Her Heritage”, Social Science Computer Review, Volume 24, Number 4, 516-523) Abstract Social interaction with computers is a key part of the frontier of automation. Weizenbaum's ELIZA continues to be relevant as the focus of the debate shifts from artificial intelligence to the future of narrative and emotional design. A version of ELIZA was implemented in Excel/Visual Basic for Applications that combines adherence to the original specification with easy experimentation. Keyword list: human computer interaction, automation, ELIZA, artificial intelligence, narrative It is a truism that automation is perpetually on the rise. But where are the limits? This has been the subject of debate for decades. One of the most lasting and highly cited contributions to the debate is ELIZA, the pioneering conversation or chatterbot program originally developed by Weizenbaum (1966). It played a role in the debate on Artificial Intelligence in the 1970s, especially as an example to show that Artificial Intelligence can hardly be more than simulated rather than emulated intelligence (Raggett and Bains 1992: 204). However, interest in ELIZA survived the AI hype. It surfaced in recent discussions on identity and on-line interaction (Turkle 1995), the future of narrative (Murray 1997) and emotional design (Norman 2004). ELIZA did not become a classic because of her technical sophistication. Kurzweil (1990: 36), for example, described her as a “simple minded program unrepresentative of the state of the art”. -
' MACSYMA Users' Conference
' MACSYMA Users'Conference Held at University of California Berkeley, California July 27-29, 1977 I TECH LIBRARY KAFB, NY NASA CP-2012 Proceedings of the 1977 MACSYMA Users’ Conference Sponsored by Massachusetts Institute of Technology, University of California at Berkeley, NASA Langley Research Center and held at Berkeley, California July 27-29, 1977 Scientific and TechnicalInformation Office 1977 NATIONALAERONAUTICS AND SPACE ADMINISTRATION NA5A Washington, D.C. FOREWORD The technical programof the 1977 MACSPMA Users' Conference, held at Berkeley,California, from July 27 to July 29, 1977, consisted of the 45 contributedpapers reported in.this publicationand of a workshop.The work- shop was designed to promote an exchange of information between implementers and users of the MACSYMA computersystem and to help guide future developments. I The response to the call for papers has well exceeded the early estimates of the conference organizers; and the high quality and broad ra.ngeof topics of thepapers submitted has been most satisfying. A bibliography of papers concerned with the MACSYMA system is included at the endof this publication. We would like to thank the members of the programcommittee, the many referees, and the secretarial and technical staffs at the University of California at Berkeley and at the Laboratory for Computer Science, Massachusetts Instituteof Technology, for shepherding the many papersthrough the submission- to-publicationprocess. We are especiallyappreciative of theburden. carried by .V. Ellen Lewis of M. I. T. for serving as expert in document preparation from computer-readableto camera-ready copy for several papers. This conference originated as the result of an organizing session called by Joel Moses of M.I.T. -
From Chatbot to Personal Assistant
6e conférence conjointe Journées d'Études sur la Parole (JEP, 31e édition), Traitement Automatique des Langues Naturelles (TALN, 27e édition), Rencontre des Étudiants Chercheurs en Informatique pour le Traitement Automatique des Langues (RÉCITAL, 22e édition) Nancy, France, 08-19 juin 2020 From chatbot to personal assistant. Emma JEANNETTE Sawssen HADDED Université de Paris, 5 Rue Thomas Mann, 75013 Paris, France [email protected], [email protected] RÉSUMÉ Les chatbots (emprunt direct de l’anglais, chat = conversation amicale et informelle; bot = un programme informatique avec un fonctionnement automatique) sont des logiciels informatiques créés pour simuler des conversations à ressemblance presque humanoïde. Ces logiciels analysent les textes ou commandes vocales pour adapter leurs réponses. Eliza est un des tout premiers chatbots crée par Weizenbaum entre 1964 et 1966. Le but initial d’Eliza était de démontrer que la communication entre les humains et les machines n’était pas nécessaire. Seulement le programme a engendré de nouvelles recherches sur les machines et le langage est devenu l’ancêtre de nos chatbots actuels. Notre expérience se concentre sur les différents chatbots de renommée au fil des années de 1964 à 2018, leurs améliorations mais aussi leurs limitations. Nous avons testés chaque chatbot (ELIZA,A.L.I.C.E. 1995,Cleverbot 1997, Mitusku 2005) sur des phrases tests. Phrases de protocole 1 Tout d’abord, nous nous sommes présentées sous un faux nom et nous leur avons demandé de répéter ce nom plus tard dans la conversation pour voir s'ils enregistraient les informations sur l’utilisateur. Puis une question basique (a) pour voir s'ils pouvaient répondre à des questions simples. -
Riddle Machines: the History and Nature of Interactive Fiction
Riddle Machines: The History and Nature of Interactive Fiction The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation Montfort, Nick. "Riddle Machines: The History and Nature of Interactive Fiction." A Companion to Digital Literary Studies, edited by Ray Siemens and Susan Schreibman, John Wiley & Sons Ltd, 2013, 267-282. © 2013 Ray Siemens and Susan Schreibman As Published http://dx.doi.org/10.1002/9781405177504.ch14 Publisher John Wiley & Sons, Ltd Version Author's final manuscript Citable link https://hdl.handle.net/1721.1/129076 Terms of Use Creative Commons Attribution-Noncommercial-Share Alike Detailed Terms http://creativecommons.org/licenses/by-nc-sa/4.0/ Nick Montfort Riddle Machines: The History and Nature of Interactive Fiction 14. Riddle Machines: The History and Nature of Interactive Fiction Nick Montfort Introduction The genre that has also been labeled "text adventure" and "text game" is stereotypically thought to offer dungeons, dragons, and the ability for readers to choose their own adventure. While there may be dragons here, interactive fiction (abbreviated "IF") also offers utopias, revenge plays, horrors, parables, intrigues, and codework, and pieces in this form resound with and rework Gilgamesh, Shakespeare, and Eliot as well as Tolkien. The reader types in phrases to participate in a dialogue with the system, commanding a character with writing. Beneath this surface conversation, and determining what the computer narrates, there is the machinery of a simulated world, capable of drawing the reader into imagining new perspectives and understanding strange systems. Interactive fiction works can be challenging for literary readers, even those interested in other sorts of electronic literature, because of the text-based interface and because of the way in which these works require detailed exploration, mapping, and solution. -
Edward Feigenbaum 1994 ACM Turing Award Recipient
Edward Feigenbaum 1994 ACM Turing Award Recipient Interviewed by: Don Knuth Recorded: April 4, and May 2, 2007 Mountain View, California This transcript, and the video interview on which it is based, is copyright by the Computer History Museum and the ACM is using it by their generous permission. © 2007 Computer History Museum CHM Reference number: X3897.2007 DK: Don Knuth, the interviewer EF: Edward (Ed) Feigenbaum, the 1994 ACM Turing Award Recipient DK: Hello, everyone. My name is Don Knuth. Today [April 4, 2007] I have the great privilege e of interviewing Ed Feigenbaum for oral history where we hope to reach people generations from now and also people of today. I’m going to try to ask a zillion questions about things that I wish I could have asked people who unfortunately are dead now so that students and general scientific people, people with general interest in the world, historians and so on in the future will have some idea as to what Ed was really like even though none of us are going to be around forever. Now this is the second part of a two- phase study. Ed grilled me a couple weeks ago and now it’s my turn to do him. I hope I can do half as well as he did for me. I’d like to get right into it -- But first of all sort of for organization, my plan -- who knows if it’ll carry it out -- is to do a little bit first that’s chronological, in order to set the scenes for his whole life. -
From Eliza to Xiaoice: Challenges and Opportunities with Social Chatbots
From Eliza to XiaoIce: Challenges and Opportunities with Social Chatbots Heung-Yeung Shum, Xiaodong He, Di Li Microsoft Corporation Abstract: Conversational systems have come a long way since their inception in the 1960s. After decades of research and development, we’ve seen progress from Eliza and Parry in the 60’s and 70’s, to task-completion systems as in the DARPA Communicator program in the 2000s, to intelligent personal assistants such as Siri in the 2010s, to today’s social chatbots like XiaoIce. Social chatbots’ appeal lies not only in their ability to respond to users’ diverse requests, but also in being able to establish an emotional connection with users. The latter is done by satisfying users’ need for communication, affection, as well as social belonging. To further the advancement and adoption of social chatbots, their design must focus on user engagement and take both intellectual quotient (IQ) and emotional quotient (EQ) into account. Users should want to engage with a social chatbot; as such, we define the success metric for social chatbots as conversation-turns per session (CPS). Using XiaoIce as an illustrative example, we discuss key technologies in building social chatbots from core chat to visual awareness to skills. We also show how XiaoIce can dynamically recognize emotion and engage the user throughout long conversations with appropriate interpersonal responses. As we become the first generation of humans ever living with AI, we have a responsibility to design social chatbots to be both useful and empathetic, so they will become ubiquitous and help society as a whole. 1. -
Chatbots: History, Technology, and a Case Analysis Jay 1
Chatbots: history, technology, and a case analysis Item Type Honor's Project Authors Jay, Benjamin Rights Attribution-NonCommercial-NoDerivatives 4.0 International Download date 27/09/2021 10:16:35 Item License http://creativecommons.org/licenses/by-nc-nd/4.0/ Link to Item http://hdl.handle.net/20.500.12648/1624 Running head: Chatbots: History, Technology, and a Case Analysis Jay 1 Chatbots: History, Technology, and a Case Analysis Benjamin Jay State University of New York at New Paltz Chatbots: History, Technology, and a Case Analysis Jay 2 Abstract This thesis examines the more than 50 year history of chatbots that led to the development of Amazon’s Alexa, Google’s Assistant, and Apple’s Siri. A chatbot, commonly known as a conversational agent, is a computer framework that can have a normal conversation with a user by using a natural language processor (Reshmi and Balakrishnan, 2018). The goal is to understand the psychological and mathematical theories that worked well throughout history, as well as those that did not, and the impact they had on the evolution of modern chatbots. This thesis incorporates these theories into a new chatbot created using Google’s chatbot AI platform called Dialogflow. By following a Coursera course titled Building Conversational Experiences with Dialogflow, this thesis creates a chatbot that can schedule tours of a school and can answer questions about the SUNY New Paltz 2020 Commencement ceremony. Creating even the most basic chatbot requires a comprehensive understanding of the underlying theories and extensive coding experience (Abdul-Kader & Woods, 2015). This thesis assumes a foundation knowledge of computer coding. -
Chatbots' Greetings to Human-Computer Communication
Chatbots’ Greetings to Human-Computer Communication Maria Joao˜ Pereira1, Lu´ısa Coheur1;2, Pedro Fialho1, Ricardo Ribeiro1;3 1INESC-ID Lisboa, 2IST, Universidade de Lisboa, 3Instituto Universitario´ de Lisboa (ISCTE-IUL) Rua Alvez Redol, 9, 1000-029 Lisboa, Portugal [email protected], [email protected], peter.fi[email protected], [email protected] Abstract In the last years, chatbots have gained new attention, due to the interest showed by widely known personalities and companies. The concept is broad, and, in this paper we target the work developed by the (old) community that is typically associated with chatbot’s competitions. In our opinion, they contribute with very interesting know-how, but specially with large-scale corpora, gathered by interactions with real people, an invaluable resource considering the renewed interest in Deep Nets. Keywords: natural language interfaces, agent-based interaction, intelligent agents, interaction design 1. Introduction of references to web pages in this paper) makes it difficult Chatbots are currently a hot-topic, both for industry and to uncover the technology and the real possibilities behind academia (Dale, 2016; Følstad and Brandtzæg, 2017). them. In this paper, we unveil the main contributions of There are many platforms to help developing such systems, this community, as we believe that this line of work can and the number of new chatbots continues to increase at bring important insights to the human-machine communi- a dizzying pace. Pandorabots hosting service1 declares to cation field, as some of them contribute with large amounts have more than 225,000 botmasters (people in charge of of data gathered during their interactions with the crowd, creating/maintaining a chatbot), which have built more than which could be used by current data-driven chatbots (e.g., 280,000 chatbots, resulting in more than 3 billion interac- (Li et al., 2016; Vinyals and Le, 2015)).