Building University-Industry Relations in the Computer Science Department at Stanford University 1963-1972

Total Page:16

File Type:pdf, Size:1020Kb

Building University-Industry Relations in the Computer Science Department at Stanford University 1963-1972 Title: Academic Revolution and Regional Innovation: Building University-Industry Relations in the Computer Science Department at Stanford University 1963-1972 Author: Danny Crichton is beginning a Fulbright fellowship as a visiting researcher at KAIST in South Korea. Before beginning this position, he was an undergraduate at Stanford University, where he majored in Mathematical and Computational Sciences with honors in Science, Technology, and Society. His senior thesis was sponsored by a grant from the university, and he won Stanford’s Firestone Award for Undergraduate Research. He can be contacted at [email protected] Subtheme: Silicon Valley – Exploring the ‘works of the engine’, History and Conditions for success Keywords: Stanford, Computer Science, History, Academic Entrepreneurship, University- Industry Relations Copyright of the paper belongs to the author(s). Submission of a paper grants permission to the Triple Helix 9 Scientific Committee to include it in the conference material and to place it on relevant websites. The Scientific Committee may invite accepted papers accepted to be considered for publication in Special Issues of selected journals after the conference. Introduction There remains little consensus in regional studies on the origins of Silicon Valley or other innovation hubs. Different approaches, reflecting the interdisciplinary nature of the field, have examined the issue from institutional, cultural, and network analy- sis perspectives. At the same time, historians of science are beginning to construct a more detailed narrative of the development of computer science in the United States, particularly in the divide between academic theory and industrial practice. Akera has developed a theoretically rigorous synthesis of the field’s rise as part of his study on the pluralism of computation in the Cold War era. He develops the notion of an ecology of knowledge to show how the tension between military applications, com- mercial goals and academic desires shaped the direction of computing. On university campuses, this tension was manifest between the academic staff of the discipline and the service staff of university computational facilities, a conflict that eventually led to their “disintegration.”1 Looking at the people of computer science, Ensmenger writes that the need for aca- demic legitimacy was a crucial element in the direction of computer science depart- ments, and this concern caused departments to focus on theoretical concepts (especially the algorithm) as a means of building a defined field of inquiry with open problems and clear research directions. He argues that this increasing theoretical basis assisted aca- 1Atsushi Akera, Calculating a Natural World, MIT Press, 2008. 1 demic departments, but led to a widening gap between the science and the applications of computer science.2 However, the models developed by these studies are incongruent with the experi- ence at Stanford University. After joining the Mathematics department in 1957, George E. Forsythe, a professor of numerical analysis, worked to quickly develop computer science into its own discipline. Computer science was soon provided its own division within the Mathematics department, allowing Forsythe a level of independence for the burgeoning area of study. These early years were tough for the Computer Science division. Faculty billets were shared with the Mathematics department, ensuring a constant friction over staff. Furthermore, the development of computer science as a discipline brought its research program away from the work conducted by other mathematicians, generating impor- tant discussions on the utility and legitimacy of this new discipline. Eventually, these disagreements would cause Forsythe and the Computer Science division to leave the Mathematics department and create an independent division at the end of 1963. Despite these issues of academic legitimacy, the division’s growth led to the univer- sity administration granting full department status on January 1, 1965. However, the financial pressures on the department continued to grow as high inflation and university budget cuts constrained its expansion. Securing fungible funds was challenging, and 2Nathan Ensmenger, The Computer Boys Take Over: Computers, Programmers, and the Politics of Technical Expertise, MIT Press, 2010. 2 the department could not rely exclusively on the Stanford administration and the federal government to provide the funds required to meet student and research demands. Un- like the MIT and Michigan experiences documented by Akera, a relative peace between the service and academic wings was typical at Stanford throughout this period, and the Computation Center actively subsidized the academic Computer Science department. More importantly, the Computer Science department began to look outside of the university and the federal government for funds. Despite its theoretical character, the department would begin a series of programs to engage industry and develop new and durable revenue sources, in contrast to the picture developed by Ensmenger. Providing a model to frame this analysis, Etzkovitz and Leydesdorff have developed the theory of the “triple helix” to describe the relations between universities, industry and the government. While the three types of institutions are generally described as being part of a triangle, the triple helix model takes as a basis the differential approaches of the three groups and adds elements of co-evolution (generating the ever-evolving helix).3 Etzkovitz has further developed these notions in analyzing the development of “entrepreneurial science” at MIT and Stanford.4 This paper investigates the role of industry in the Computer Science programs at Stanford, and how the Computer Science department adapted to engage industry. It ar- gues that the Computer Science department created venues of engagement that allowed 3Universities and the Global Knowledge Economy: A Triple Helix of University-Industry-Government Relations, Ed. Henry Etzkovitz and Loet Leydesdorff, Pinter, 1997 4Henry Etzkovitz, MIT and the Rise of Entrepreneurial Science, Routledge, 2002. 3 for the circulation of talent that spread crucial ideas between the university and industry. Along the way, these relations provided direct financial benefits to the department, either in the form of money or equipment. Thus, the growth of university-industry relations served several different purposes for each side of the partnerships. This paper begins by looking at the development of the Computer Science depart- ment and the conflict over its academic legitimacy. Then, a brief history of the financial situation of the program’s construction of a new building provides a framework for un- derstanding the department’s need to secure industry partners. The remainder of the pa- per analyzes the development of industry grant sources, particularly from IBM, as well as the development of venues to engage industry such as the Honors Co-op program, which provided a means for industry engineers to take classes at Stanford conveniently, and the Computer Forum, a conference for faculty and top industry engineers to meet and discuss research. Developing Computer Science George E. Forsythe joined the Stanford University Mathematics department in 1957 as a full professor, joining John Herriot as a numerical analyst. Herriot was among the first of the leaders of computation at Stanford, and the two men immediately began to consider approaches for developing the field of computer science on an educational and intellectual level within the department. 4 In the years before 1961, there was no official administrative structure for computer science, although Forsythe did formulate a sub-field of sorts within the Mathematics department by 1959. All decisions regarding the academic side of computer science in these years were passed through David Gilbarg, the chair of the department. Gilbarg’s interests were in algebraic number theory early in his career, but his work in World War II led him to focus on nonlinear partial differential equations and fluid dynamics for the remainder of his career. When Forsythe arrived, the differences between the area now defined as computer science and the traditional field of mathematics were relatively few. Mathematics hired Forsythe to add strength in numerical analysis, and he strongly believed in the utility of numerical approaches, which he considered to be the next stage in the development of mathematics. He urged his colleagues that a mathematics education should include at least a basic background in using computation.5 The university administration was also enthusiastic about the new field. Albert Bowker, a statistician who was an associate dean in the School of Humanities and Sci- ences (H&S), discussed the formation of an autonomous division for the field from the very start of its formation within Mathematics.6 By 1961, the discussion had moved to the issue of logistics, and how such a division would be formed and operated within 5Donald E. Knuth, “George Forsythe and the Development of Computer Science,” Communications of the ACM, Vol. 15, No. 8 (Aug. 1972), pg. 721-727. 6The idea of an autonomous division within Mathematics was a bureaucratic construction. The basic design was to place Forsythe as head of the division with his own budget, but with the chair of Mathe- matics continuing to hold final administrative authority. The change allowed Stanford to argue that it had an autonomous Computer Science unit. 5 H&S. Gilbarg was not actively a
Recommended publications
  • The Ethics of Machine Learning
    SINTEZA 2019 INTERNATIONAL SCIENTIFIC CONFERENCE ON INFORMATION TECHNOLOGY AND DATA RELATED RESEARCH DATA SCIENCE & DIGITAL BROADCASTING SYSTEMS THE ETHICS OF MACHINE LEARNING Danica Simić* Abstract: Data science and machine learning are advancing at a fast pace, which is why Nebojša Bačanin Džakula tech industry needs to keep up with their latest trends. This paper illustrates how artificial intelligence and automation in particular can be used to en- Singidunum University, hance our lives, by improving our productivity and assisting us at our work. Belgrade, Serbia However, wrong comprehension and use of the aforementioned techniques could have catastrophic consequences. The paper will introduce readers to the terms of artificial intelligence, data science and machine learning and review one of the most recent language models developed by non-profit organization OpenAI. The review highlights both pros and cons of the language model, predicting measures humanity can take to present artificial intelligence and automatization as models of the future. Keywords: Data Science, Machine Learning, GPT-2, Artificial Intelligence, OpenAI. 1. INTRODUCTION Th e rapid development of technology and artifi cial intelligence has al- lowed us to automate a lot of things on the internet to make our lives eas- ier. With machine learning algorithms, technology can vastly contribute to diff erent disciplines, such as marketing, medicine, sports, astronomy and physics and more. But, what if the advantages of machine learning and artifi cial intelligence in particular also carry some risks which could have severe consequences for humanity. 2. MACHINE LEARNING Defi nition Machine learning is a fi eld of data science which uses algorithms to improve the analytical model building [1] Machine learning uses artifi cial Correspondence: intelligence (AI) to learn from great data sets and identify patterns which Danica Simić could support decision making parties, with minimal human intervention.
    [Show full text]
  • Supervised Machine Learning
    Tikaram Acharya Supervised Machine Learning Accuracy Comparison of Different Algorithms on Sample Data Metropolia University of Applied Sciences Bachelor of Engineering Information Technology Bachelor’s Thesis 11 January 2021 Author Tikaram Acharya Title Supervised Machine Learning: Accuracy Comparison of Differ- ent Algorithms on Sample Data Number of Pages 39 pages Date 11 January 2021 Degree Bachelor of Engineering Degree Programme Information Technology Professional Major Software Engineering Instructors Janne Salonen, Head of School The purpose of the final year project was to discuss the theoretical concepts of the analyt- ical features of supervised machine learning. The concepts were illustrated with sample data selected from the open-source platform. The primary objectives of the thesis include explaining the concept of machine learning, presenting supportive libraries, and demon- strating their application or utilization on real data. Finally, the aim was to use machine learning to determine whether a person would pay back their loan or not. The problem was taken as a classification problem. To accomplish the goals, data was collected from an open data source community called Kaggle. The data was downloaded and processed with Python. The complete practical part or machine learning workflow was done in Python and Python’s libraries such as NumPy, Pandas, Matplotlib, Seaborn, and SciPy. For the given problem on sample data, five predictive models were constructed and tested with machine learning with different techniques and combinations. The results were highly accurately matched with the classification problem after evaluating their f1-score, confu- sion matrix, and Jaccard similarity score. Indeed, it is clear that the project achieved the main defined objectives, and the necessary architecture has been set up to apply various analytical abilities into the project.
    [Show full text]
  • Seinfeld's Recent Comments on Zoom…. Energy
    Stanford Data Science Departmental 2/4/2021 Seminar Biomedical AI: Its Roots, Evolution, and Early Days at Stanford Edward H. Shortliffe, MD, PhD Chair Emeritus and Adjunct Professor Department of Biomedical Informatics Columbia University Department of Biomedical Data Science Seminar Stanford Medicine Stanford, California 2:30pm – 4:50pm – February 4, 2021 Seinfeld’s recent comments on Zoom…. Energy, attitude and personality cannot be “remoted” through even the best fiber optic lines Jerry Seinfeld: So You Think New York Is ‘Dead’ (It’s not.) Aug. 24, 2020 © E.H. Shortliffe 1 Stanford Data Science Departmental 2/4/2021 Seminar Disclosures • No conflicts of interest with content of presentation • Offering a historical perspective on medical AI, with an emphasis on US activities in the early days and specific work I know personally • My research career was intense until I became a medical school dean in 2007 • Current research involvement is largely as a textbook author and two decades as editor-in-chief of a peer- reviewed journal (Journal of Biomedical Informatics [Elsevier]) Goals for Today’s Presentation • Show that today’s state of the art in medical AI is part of a 50-year scientific trajectory that is still evolving • Provide some advice to today’s biomedical AI researchers, drawing on that experience • Summarize where we are today in the evolution, with suggestions for future emphasis More details on slides than I can include in talk itself – Full deck available after the talk © E.H. Shortliffe 2 Stanford Data Science Departmental
    [Show full text]
  • Artificial Intelligence Machine Learning and Deep Learning
    ARTIFICIAL INTELLIGENCE MACHINE LEARNING AND DEEP LEARNING LICENSE, DISCLAIMER OF LIABILITY, AND LIMITED WARRANTY By purchasing or using this book and its companion files (the “Work”), you agree that this license grants permission to use the contents contained herein, but does not give you the right of ownership to any of the textual content in the book or ownership to any of the information, files, or products contained in it. This license does not permit uploading of the Work onto the Internet or on a network (of any kind) without the written consent of the Publisher. Duplication or dissemination of any text, code, simu- lations, images, etc. contained herein is limited to and subject to licensing terms for the respective products, and permission must be obtained from the Publisher or the owner of the content, etc., in order to reproduce or network any portion of the textual material (in any media) that is contained in the Work. MERCURY LEARNING AND INFORMATION (“MLI” or “the Publisher”) and anyone involved in the creation, writing, production, accompanying algorithms, code, or computer programs (“the software”), and any accompanying Web site or software of the Work, cannot and do not warrant the performance or results that might be obtained by using the contents of the Work. The author, developers, and the Publisher have used their best efforts to insure the accuracy and functionality of the textual material and/or pro- grams contained in this package; we, however, make no warranty of any kind, express or implied, regarding the performance of these contents or programs. The Work is sold “as is” without warranty (except for defective materials used in manufacturing the book or due to faulty workmanship).
    [Show full text]
  • The 1970 United States Computer Chess Championship: the Start of the Longest-Running Experiment in Computer Science History
    The 1970 United States Computer Chess Championship: The Start of the Longest-Running Experiment in Computer Science History Jonathan Schaeffer University of Alberta Abstract On August 31, 1970, an experiment began that continues to this day. The first chess tournament for computers was held as part of the Association for Computing Machinery’s (ACM’s) National Conference. The interest generated was tremendous, leading to ACM sponsoring an annual event until 1994. Chess competitions continue to this day, allowing for 50 years of data on the growth of artificial intelligence capabilities in this domain. During this period, program ratings have soared from roughly 1400 in 1970 to over 3500 today. The 1970 event was the first continuous competition in computer science history, and it represents the longest ongoing experiment in computer science history.1 Introduction Creating a program capable of competing with the human world chess champion was one of the first “grand challenge” problems of the fledgling research area of artificial intelligence. The importance of computer chess at the dawn of the computer age is reflected in the many computer science luminaries that contributed to the early advances in the field. This includes Claude Shannon (father of information theory), Alan Turning (creator of the Turing Machine and the Turing Test), Herbert Simon (Nobel Laureate and winner of the prestigious Turing Award), Alan Newell (Turing Award winner), and John McCarthy (Turing Award winner). In the 1950s and the 1960s, progress in developing chess programs was slow. However a breakthrough occurred with the development of MACHACK VI by Massachusetts Institute of Technology (MIT) student Richard Greenblatt.
    [Show full text]
  • AI Watch Historical Evolution of Artificial Intelligence
    JRC TECHNICAL REPORTS AI Watch Historical Evolution of Artificial Intelligence Analysis of the three main paradigm shifts in AI EUR 30221 EN This publication is a Technical report by the Joint Research Centre (JRC), the European Commission’s science and knowledge service. It aims to provide evidence-based scientific support to the European policymaking process. The scientific output expressed does not imply a policy position of the European Commission. Neither the European Commission nor any person acting on behalf of the Commission is responsible for the use that might be made of this publication. Contact information Blagoj DELIPETREV European Commission, Joint Research Centre, TP262, Via Fermi, 21027 Ispra (VA), ITALY [email protected] Tel.: + 39 0332 786 352 AI-Watch functional mailbox [email protected] EU Science Hub https://ec.europa.eu/jrc JRC120469 EUR 30221 EN PDF ISBN 978-92-76-18940-4 ISSN 1831-9424 doi:10.2760/801580 Luxembourg: Publications Office of the European Union, 2020 © European Union, 2020 The reuse policy of the European Commission is implemented by the Commission Decision 2011/833/EU of 12 December 2011 on the reuse of Commission documents (OJ L 330, 14.12.2011, p. 39). Except otherwise noted, the reuse of this document is authorised under the Creative Commons Attribution 4.0 International (CC BY 4.0) licence (https://creativecommons.org/licenses/by/4.0/). This means that reuse is allowed provided appropriate credit is given and any changes are indicated. For any use or reproduction of photos or other material that is not owned by the EU, permission must be sought directly from the copyright holders.
    [Show full text]
  • Bounded Rationality and Artificial Intelligence
    No. 2019 - 04 The Game of Go: Bounded Rationality and Artificial Intelligence Cassey Lee ISEAS – Yusof Ishak Institute [email protected] May 2019 Abstract The goal of this essay is to examine the nature and relationship between bounded rationality and artificial intelligence (AI) in the context of recent developments in the application of AI to two-player zero sum games with perfect information such as Go. This is undertaken by examining the evolution of AI programs for playing Go. Given that bounded rationality is inextricably linked to the nature of the problem to be solved, the complexity of Go is examined using cellular automata (CA). Keywords: Bounded Rationality, Artificial Intelligence, Go JEL Classification: B41, C63, C70 The Game of Go: Bounded Rationality and Artificial Intelligence Cassey Lee 1 “The rules of Go are so elegant, organic, and rigorously logical that if intelligent life forms exist elsewhere in the universe, they almost certainly play Go.” - Edward Lasker “We have never actually seen a real alien, of course. But what might they look like? My research suggests that the most intelligent aliens will actually be some form of artificial intelligence (or AI).” - Susan Schneider 1. Introduction The concepts of bounded rationality and artificial intelligence (AI) do not often appear simultaneously in research articles. Economists discussing bounded rationality rarely mention AI. This is reciprocated by computer scientists writing on AI. This is surprising because Herbert Simon was a pioneer in both bounded rationality and AI. For his contributions to these areas, he is the only person who has received the highest accolades in both economics (the Nobel Prize in 1978) and computer science (the Turing Award in 1975).2 Simon undertook research on both bounded rationality and AI simultaneously in the 1950s and these interests persisted throughout his research career.
    [Show full text]
  • TEX: a Branch in Desktop Publishing Evolution, Part 1
    IEEE ANNALS OF THE HISTORY OF COMPUTING THEME ARTICLE: HISTORY OF DESKTOP PUBLISHING, PART 1 TEX: A Branch in Desktop Publishing Evolution, Part 1 Donald Knuth began the development of TEX in 1977 and had an initial version running in 1978, with the Barbara Beeton American Mathematical Society aim of typesetting mathematical documents with the Karl Berry highest quality, and with archival reproducibility far TEX Users Group into the future. Its usage spread, and today the TEX David Walden system remains in use by a vibrant user community. However, the world of TEX has always been somewhat out of sync with the commercial desktop publishing systems that began to come out in the mid-1980s and are now ubiquitous in the worlds of publishing and printing. TEX is a typesetting system that was developed by Donald Knuth, starting in 1977, to “do beau- tiful typesetting,” particularly of mathematics. At the time, PostScript and PDF did not yet exist, people were in the early stages of mathematically defining digital fonts, and WYSIWYG user interfaces were just being conceived. What we today think of as desktop publishing systems (for example, WYSIWYG, PostScript output) were half a dozen years away, and the computers of the time were slow and small compared with today. Thus, TEX was implemented using markup of plain text files, as was typical with prior systems such as RUNOFF, to instruct TEX about docu- ment formatting. TEX also had its own way of outputting page descriptions for print as well as its own companion system, Metafont, for designing and formatting type.
    [Show full text]
  • Collective Test
    COLLECTIVE TEST: CAN GENERATIVE TOOLS WORKING AUTONOMOUSLY CREATE NEW CULTURES IN THE LONG RUN? Architectural history thesis spring 2021 spring 2021 AR2A011 ARCHITECTURAL HISTORY THESIS Students: Haruka Maeda Leticija Aleksandra Petrova Tutor: Dan C. Baciu COLLECTIVE TEST: CAN GENERATIVE TOOLS WORKING AUTONOMOUSLY CREATE NEW CULTURES IN THE LONG RUN? Architectural history thesis spring 2021 Index Abstract 00 Introduction 01 -Brief history of AI and generative tools -Current testing of generative tools& limitation -Neccessity for an additional test, breaking from limitations -Methodology -The necessary steps -The choice of the 3 types of case studies -The specific projects & the tools used Generative tools: 02 created Illusion and human trace -Introduction for case studies -Approach of a human designer -Case study 1: AI & Text Illusion and the turing test The process and human trace -Case study 2: AI & Art Illusion and the turing test The process and human trace -Case study 3: AI & Architecture Illusion and the turing test The process and human trace Summary of the created illusions and human inputs Can generative tools 03 autonomously create human cultures on a larger scale? Conclusion: Collective test 04 Bibliography 05 Appendix 06 Abstract The term AI has entered the creative industry in the last decades. Computer based design is now ominiprescent in everyday life. Will AI take over the creative industry? Can they autonomously be creative? These are some of the headlines propagated by the mainstream media. This paper will unveil the human labor behind these computer generated designs by analyzing how they were publicized and how they were actually developed. Through the related case studies, it is possible to identify how much of the AI work is still human originated, and how a total automation is still not attained.
    [Show full text]
  • John Von Neumann
    John von Neumann ! Hungarian mathematician, computer scientist, cyberneticist, all-around genius ! Worked on atomic bomb project in WW II ! Invented game theory and developed theory of self-replicating automata ! Originated key concept of stored-program computer in 1945 ! Program instructions = data ! Easily reprogrammable ! Von Neumann architecture is still the universal standard EDVAC ! Electronic Discrete Variable Automatic Computer ! Designed by Mauchly, Eckert, and Von Neumann ! Stored-program design ! Used binary instead of decimal to represent information ! Version called UNIVAC I was the first commercially available computer system ! Sold to the U.S. Census Bureau in 1951 First Generation Computers ! Mid 1940s to late 1950s ! Stored-program design with ~ 1000 words of RAM ! Used vacuum tubes, but required less space than ENIAC ! Punched cards for input and output ! Vacuum tube or magnetic core memory for data storage ! Programmed directly in binary machine language ! Included EDVAC and UNIVAC First Generation Computers Transistors ! Invented at Bell Labs in 1947 by William Shockley, John Bardeen, and Walter Brattain ! Generated far less heat than vacuum tubes ! Required far less power ! Much faster, smaller, cheaper, and more reliable Transistors ! Incorporated into Second Generation computers in the late 1950s and early 1960s Origins of AI ! 1956 Summer Dartmouth Conference – Organized by John McCarthy, who coined the term “artificial intelligence” – Participants: ! John McCarthy ! Marvin Minsky ! Claude Shannon Minsky Shannon
    [Show full text]
  • In Memoriam-Arthur Samuel: Pioneer in Machine Learning
    Memoriam AI Magazine Volume 11 Number 3 (1990) (© AAAI) In Memoriam Arthur Samuel: Pioneer in Samuel was a modest man, and the strate the Machine Learning importance of his work was widely power of recognized only after his retirement electronic Arthur Samuel (1901–1990) was a from IBM in 1966, in part because he comput- pioneer of artificial intelligence didn’t relish the politics that were ers. He research. From 1949 through the late required to have his research more didn’t fin- 1960s, he did the best work in mak- vigorously followed up on. He was ish the ing computers learn from their expe- also realistic about the large difference program while he was at the universi- rience. His vehicle for this work was between what had been accomplished ty of Illinois, perhaps because the the game of checkers. in understanding intellectual mecha- computer wasn’t finished in time. Programs for playing games often nisms and what would be required to In 1949, Samuel joined IBM’s fill the role in artificial intelligence reach human-level intelligence. Poughkeepsie Laboratory, where he research that the fruit fly Drosophila Samuel’s papers on machine learn- worked on IBM’s first stored program plays in genetics. Drosophilae are ing are still worth studying. Possess- computer, the 701. This computer’s convenient for genetics because they ing great creativity and essentially memory was based on Williams tubes, breed fast and are cheap to keep, and working alone, doing his own pro- which stored bits as charged spots on games are convenient for artificial gramming, he invented several semi- the screen of a cathode ray tube.
    [Show full text]
  • Introduction
    15-780: Graduate AI Introduction Instructors: Tuomas Sandholm and Manuela Veloso TAs: John Dickerson and Prateek Tandon http://www.cs.cmu.edu/~15780 Carnegie Mellon University Evaluation n 4 Homeworks 40% n Midterm Test 20% n Final Project 30% n Class participation 10% n Homeworks are done individually n 8 total “mercy” days (no penalty) for late homeworks - cannot use more than 3 mercy days in a single homework. No credit for late homeworks with no mercy days. n Final projects can be done in groups of up to 2 people 1 Resources n Lectures n Presentation and discussion in class n Lecture slides rd n Textbook: Russell and Norvig, 3 edition n Instructors – office hours by appointment n TAs – office hours will be announced More Information n TAs will hold recitations as needed n Doodle poll for best time n Audits n Register by fill out audit form n Must do final project n Waitlist n Let us know if you are on it n Mailing list n 15780students @ cs.cmu.edu n We will send a test message 2 What is Artificial Intelligence? What is “intelligence” ? Can we emulate intelligent behavior in machines ? How far can we take it ? Time Line (NYTimes 2011) 3 n Performed for 84 years n Defeated Napoleon and Franklin n Amazon Mechanical Turk: “artificial artificial intelligence” 7 Amazon Mechanical Turk 8 4 n Can a general machine think? n Can machines “pass” the imitation game? n Turing Test: Evaluator communicates via text channel with computer and human, must reliably identify the computer 9 Views of AI Think like humans Think rationally Cognitive Science Formalize inference into laws of thought Act like humans Act rationally Turing test Act according to laws 5 The Dartmouth Conference “We propose that a two-month, ten-man study of artificial intelligence be carried out during the summer of 1956 at Dartmouth College in Hanover, NH.
    [Show full text]