History of Computers 1 Sources

Total Page:16

File Type:pdf, Size:1020Kb

History of Computers 1 Sources History of Computers The abacus, a simple counting aid, may have been invented in Babylonia (now Iraq) in the fourth century B.C. The Antikythera mechanism, used for registering and predicting the motion of the stars and planets, is dated to the first century B.C. It was discovered off the coast of Greece in 1901. Arabic numerals are introduced to Europe in the eighth and ninth centuries A.D. Roman numerals remain in use in some parts of Europe until the seventeenth century. The Arabic system introduced the concepts of the zero and fixed places for tens, hundreds, thousand, etc., and greatly simplified mathematical calculations. John Napier, Baron of Merchiston, Scotland, invents logs in 1614. Logs allow multiplication and division to be reduced to addition and subtraction. Wilhelm Schickard builds the first mechanical calculator in 1623. It can work with six digits, and carries digits across columns. It works, but never makes it beyond the prototype stage. Schickard is a professor at the University of Tubingen, Germany. Blaise Pascal builds a mechanical calculator in 1642. It has the capacity for eight digits, but has trouble carrying and its gears tend to jam. Joseph-Marie Jacquard invents an automatic loom controlled by punch cards. Charles Babbage conceives of a "Difference Engine" in 1820 or 1821. It is a massive steam-powered mechanical calculator designed to print astronomical tables. He attempts to build it over the course of the next 20 years, only to have the project cancelled by the British government in 1842. Babbage's next idea is the Analytical Engine - a mechanical computer that can solve any mathematical problem. It uses punch-cards similar to those used by the Jacquard loom and can perform simple conditional operations. Augusta Ada Byron, the countess of Lovelace, met Babbage in 1833. She describes the Analytical Engine as weaving "algebraic patterns just as the Jacquard loom weaves flowers and leaves." Her published analysis of the Analytical Engine is our best record of its programming potential. In it she outlines the fundamentals of computer programming, including data analysis, looping and memory addressing. Konrad Zuse, a German engineer, completes the first general purpose progammable calculator in 1941. He pioneers the use of binary math and boolean logic in electronic calculation. Colossus, a British computer used for code-breaking, is operational by December of 1943. ENIAC, or Electronic Numerical Integrator Analyzor and Computer, is developed by the Ballistics Research Laboratory in Maryland to assist in the preparation of firing tables for artillery. It is built at the University of Pennsylvania's Moore School of Electrical Engineering and completed in November 1945. 1 Sources: http://www.pbs.org/nerds/timeline/index.html and http://inventors.about.com/library/blcoindex.htm History of Computers Bell Telephone Laboratories develops the transistor in 1947. UNIVAC, the Universal Automatic Computer (pictured below), is developed in 1951. It can store 12,000 digits in random access mercury-delay lines. EDVAC, for Electronic Discrete Variable Computer, is completed under contract for the Ordinance Department in 1952. In 1952 G.W. Dummer, a radar expert from the British Royal Radar Establishment, proposes that electronic equipment be manufactured as a solid block with no connecting wires. The prototype he builds doesn't work and he receives little support for his research. Texas Instruments and Fairchild semiconductor both announce the integrated circuit in 1959. The IBM 360 is introduced in April of 1964 and quickly becomes the standard institutional mainframe computer. By the mid-80s the 360 and its descendents will have generated more than $100 billion in revenue for IBM. Texas Instruments and Fairchild semiconductor both announce the integrated circuit in 1959. Ivan Sutherland demonstrates a program called Sketchpad on a TX-2 mainframe at MIT's Lincoln Labs in 1962. It allows him to make engineering drawings with a light pen. A typical minicomputer costs about $20,000. 1965: An IC that cost $1000 in 1959 now costs less than $10. Gordon Moore predicts that the number of components in an IC will double every year. This is known as Moore's Law. Doug Engelbart demonstrates in 1968 a word processor, an early hypertext system and a collaborative application: three now common computer applications. Gordon Moore and Robert Noyce found Intel in 1968. Xerox creates its Palo Alto Research Center - Xerox PARC - in 1969. Its mission is to explore the "architecture of information." Fairchild Semiconductor introduces a 256-bit RAM chip in 1970. In late 1970 Intel introduces a 1K RAM chip and the 4004, a 4-bit microprocessor. Two years later comes the 8008, an 8-bit microprocessor. Bill Gates and Paul Allen form Traf-O-Data in 1971 to sell their computer traffic-analysis systems. 1972: Gary Kildall writes PL/M, the first high-level programming language for the Intel microprocessor. 2 Sources: http://www.pbs.org/nerds/timeline/index.html and http://inventors.about.com/library/blcoindex.htm History of Computers Steve Jobs and Steve Wozniak are building and selling "blue boxes" in Southern California in 1971. April 1972: Intel introduces the 8008, the first 8-bit microprocessor. Jonathan A. Titus designs the Mark-8, "Your Personal Minicomputer," according to the July, 1974 cover of Radio- Electronics. Popular Electronics features the MITS Altair 8800 on its cover, January 1975. It is hailed as the first "personal" computer. Thousands of orders for the 8800 rescue MITS from bankruptcy. Pictured below: The Homebrew Computer Club in 1975. Paul Allen and Bill Gates develop BASIC for the Altair 8800. Microsoft is born. 1977: Apple is selling its Apple II for $1,195, including 16K of RAM but no monitor. Software Arts develops the first spreadsheet program, Visicalc, by the spring of 1979. It is released in October and is an immediate success. Copies shipped per month rise from 500 to 12,000 between 1979 and 1981. By 1980 Apple has captured 50% of the personal computer market. In 1980 Microsoft is approached by IBM to develop BASIC for its personal computer project. The IBM PC is released in August, 1981. The Apple Macintosh debuts in 1984. It features a simple, graphical interface, uses the 8-MHz, 32-bit Motorola 68000 CPU, and has a built-in 9- inch B/W screen. Microsoft Windows 1.0 ships in November, 1985. Motorola announces the 68040, a 32-bit 25MHz microprocessor. Microsoft's sales for 1989 reach $1 billion, the first year to do so. Timesharing, the concept of linking a large numbers of users to a single computer via remote terminals, is developed at MIT in the late 50s and early 60s. 1962: Paul Baran of RAND develops the idea of distributed, packet-switching networks. ARPANET goes online in 1969. Bob Kahn and Vint Cerf develop the basic ideas of the Internet in 1973. In 1974 BBN opens the first public packet-switched network - Telenet. 3 Sources: http://www.pbs.org/nerds/timeline/index.html and http://inventors.about.com/library/blcoindex.htm History of Computers A UUCP link between the University of North Carolina at Chapel Hill and Duke University establishes USENET in 1979. The first MUD is also developed in 1979, at the University of Essex. TCP/IP (Transmission Control Protocol and Internet Protocol) is established as the standard for ARPANET in 1982. 1987: the number of network hosts breaks 10,000. 1989: the number of hosts breaks 100,000. Tim Berners-Lee develops the World Wide Web. CERN releases the first Web server in 1991. 1992: the number of hosts breaks 1,000,000. The World Wide Web sports a growth rate of 341,634% in service traffic in its third year, 1993. The main U.S. Internet backbone traffic begins routing through commercial providers as NSFNET reverts to a research network in 1994. The Internet 1996 World Exposition is the first World's Fair to be held on the internet. WORLD INTERNET CONNECTIVITY (As of 6/15/95) 4 Sources: http://www.pbs.org/nerds/timeline/index.html and http://inventors.about.com/library/blcoindex.htm History of Computers "Who invented the computer?" is not a question with a simple answer. The real answer is that many inventors contributed to the history of computers and that a computer is a complex piece of machinery made up of many parts, each of which can be considered a separate invention. This series covers many of the major milestones in computer history (but not all of them) with a concentration on the history of personal home computers. Computer History Computer History Computer History Year/Enter Inventors/Inventions Description of Event 1936 Konrad Zuse - Z1 Computer First freely programmable computer. 1942 John Atanasoff & Clifford Berry Who was first in the computing biz is not ABC Computer always as easy as ABC. 1944 Howard Aiken & Grace Hopper The Harvard Mark 1 computer. Harvard Mark I Computer 1946 John Presper Eckert & John W. 20,000 vacuum tubes later... Mauchly ENIAC 1 Computer 1948 Frederic Williams & Tom Kilburn Baby and the Williams Tube turn on the Manchester Baby Computer & The memories. Williams Tube 1947/48 John Bardeen, Walter Brattain & No, a transistor is not a computer, but this Wiliam Shockley invention greatly affected the history of The Transistor computers. 1951 John Presper Eckert & John W. First commercial computer & able to pick Mauchly presidential winners. UNIVAC Computer 1953 International Business Machines IBM enters into 'The History of Computers'. IBM 701 EDPM Computer 1954 John Backus & IBM The first successful high level programming FORTRAN Computer Programming language. Language Stanford Research Institute, Bank of The first bank industry computer - also America, and General Electric MICR (magnetic ink character recognition) 1955 ERMA and MICR for reading checks.
Recommended publications
  • Download Sample
    Entrepreneurship Entrepreneurship FIFTH EDITION Andrew Zacharakis Babson College Andrew C. Corbett Babson College William D. Bygrave Babson College VP AND EDITORIAL DIRECTOR Mike McDonald EXECUTIVE EDITOR Lise Johnson EDITORIAL MANAGER Judy Howarth CONTENT MANAGEMENT DIRECTOR Lisa Wojcik CONTENT MANAGER Nichole Urban SENIOR CONTENT SPECIALIST Nicole Repasky PRODUCTION EDITOR Indirakumari, S. COVER PHOTO CREDIT © PitukTV/Shutterstock This book was set in 10/12pt Times by SPi Global, Pondicherry, India and printed and bound by Quad Graphics. Founded in 1807, John Wiley & Sons, Inc. has been a valued source of knowledge and understanding for more than 200 years, helping people around the world meet their needs and fulfill their aspirations. Our company is built on a foundation of principles that include responsibility to the communities we serve and where we live and work. In 2008, we launched a Corporate Citizenship Initiative, a global effort to address the environmental, social, economic, and ethical challenges we face in our business. Among the issues we are addressing are carbon impact, paper specifications and procurement, ethical conduct within our business and among our vendors, and community and charitable support. For more information, please visit our website: www.wiley.com/go/citizenship. Copyright © 2020, 2017, 2014, 2011, 2008 John Wiley & Sons, Inc. All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, electronic, mechanical, photocopying, recording, scanning or otherwise, except as permitted under Sections 107 or 108 of the 1976 United States Copyright Act, without either the prior written permission of the Publisher, or authorization through payment of the appropriate per‐copy fee to the Copyright Clearance Center, Inc., 222 Rosewood Drive, Danvers, MA 01923 (Web site: www.copyright.com).
    [Show full text]
  • Historical Perspective and Further Reading 162.E1
    2.21 Historical Perspective and Further Reading 162.e1 2.21 Historical Perspective and Further Reading Th is section surveys the history of in struction set architectures over time, and we give a short history of programming languages and compilers. ISAs include accumulator architectures, general-purpose register architectures, stack architectures, and a brief history of ARMv7 and the x86. We also review the controversial subjects of high-level-language computer architectures and reduced instruction set computer architectures. Th e history of programming languages includes Fortran, Lisp, Algol, C, Cobol, Pascal, Simula, Smalltalk, C+ + , and Java, and the history of compilers includes the key milestones and the pioneers who achieved them. Accumulator Architectures Hardware was precious in the earliest stored-program computers. Consequently, computer pioneers could not aff ord the number of registers found in today’s architectures. In fact, these architectures had a single register for arithmetic instructions. Since all operations would accumulate in one register, it was called the accumulator , and this style of instruction set is given the same name. For example, accumulator Archaic EDSAC in 1949 had a single accumulator. term for register. On-line Th e three-operand format of RISC-V suggests that a single register is at least two use of it as a synonym for registers shy of our needs. Having the accumulator as both a source operand and “register” is a fairly reliable indication that the user the destination of the operation fi lls part of the shortfall, but it still leaves us one has been around quite a operand short. Th at fi nal operand is found in memory.
    [Show full text]
  • John Mccarthy
    JOHN MCCARTHY: the uncommon logician of common sense Excerpt from Out of their Minds: the lives and discoveries of 15 great computer scientists by Dennis Shasha and Cathy Lazere, Copernicus Press August 23, 2004 If you want the computer to have general intelligence, the outer structure has to be common sense knowledge and reasoning. — John McCarthy When a five-year old receives a plastic toy car, she soon pushes it and beeps the horn. She realizes that she shouldn’t roll it on the dining room table or bounce it on the floor or land it on her little brother’s head. When she returns from school, she expects to find her car in more or less the same place she last put it, because she put it outside her baby brother’s reach. The reasoning is so simple that any five-year old child can understand it, yet most computers can’t. Part of the computer’s problem has to do with its lack of knowledge about day-to-day social conventions that the five-year old has learned from her parents, such as don’t scratch the furniture and don’t injure little brothers. Another part of the problem has to do with a computer’s inability to reason as we do daily, a type of reasoning that’s foreign to conventional logic and therefore to the thinking of the average computer programmer. Conventional logic uses a form of reasoning known as deduction. Deduction permits us to conclude from statements such as “All unemployed actors are waiters, ” and “ Sebastian is an unemployed actor,” the new statement that “Sebastian is a waiter.” The main virtue of deduction is that it is “sound” — if the premises hold, then so will the conclusions.
    [Show full text]
  • John Mccarthy – Father of Artificial Intelligence
    Asia Pacific Mathematics Newsletter John McCarthy – Father of Artificial Intelligence V Rajaraman Introduction I first met John McCarthy when he visited IIT, Kanpur, in 1968. During his visit he saw that our computer centre, which I was heading, had two batch processing second generation computers — an IBM 7044/1401 and an IBM 1620, both of them were being used for “production jobs”. IBM 1620 was used primarily to teach programming to all students of IIT and IBM 7044/1401 was used by research students and faculty besides a large number of guest users from several neighbouring universities and research laboratories. There was no interactive computer available for computer science and electrical engineering students to do hardware and software research. McCarthy was a great believer in the power of time-sharing computers. John McCarthy In fact one of his first important contributions was a memo he wrote in 1957 urging the Director of the MIT In this article we summarise the contributions of Computer Centre to modify the IBM 704 into a time- John McCarthy to Computer Science. Among his sharing machine [1]. He later persuaded Digital Equip- contributions are: suggesting that the best method ment Corporation (who made the first mini computers of using computers is in an interactive mode, a mode and the PDP series of computers) to design a mini in which computers become partners of users computer with a time-sharing operating system. enabling them to solve problems. This logically led to the idea of time-sharing of large computers by many users and computing becoming a utility — much like a power utility.
    [Show full text]
  • Fpgas As Components in Heterogeneous HPC Systems: Raising the Abstraction Level of Heterogeneous Programming
    FPGAs as Components in Heterogeneous HPC Systems: Raising the Abstraction Level of Heterogeneous Programming Wim Vanderbauwhede School of Computing Science University of Glasgow A trip down memory lane 80 Years ago: The Theory Turing, Alan Mathison. "On computable numbers, with an application to the Entscheidungsproblem." J. of Math 58, no. 345-363 (1936): 5. 1936: Universal machine (Alan Turing) 1936: Lambda calculus (Alonzo Church) 1936: Stored-program concept (Konrad Zuse) 1937: Church-Turing thesis 1945: The Von Neumann architecture Church, Alonzo. "A set of postulates for the foundation of logic." Annals of mathematics (1932): 346-366. 60-40 Years ago: The Foundations The first working integrated circuit, 1958. © Texas Instruments. 1957: Fortran, John Backus, IBM 1958: First IC, Jack Kilby, Texas Instruments 1965: Moore’s law 1971: First microprocessor, Texas Instruments 1972: C, Dennis Ritchie, Bell Labs 1977: Fortran-77 1977: von Neumann bottleneck, John Backus 30 Years ago: HDLs and FPGAs Algotronix CAL1024 FPGA, 1989. © Algotronix 1984: Verilog 1984: First reprogrammable logic device, Altera 1985: First FPGA,Xilinx 1987: VHDL Standard IEEE 1076-1987 1989: Algotronix CAL1024, the first FPGA to offer random access to its control memory 20 Years ago: High-level Synthesis Page, Ian. "Closing the gap between hardware and software: hardware-software cosynthesis at Oxford." (1996): 2-2. 1996: Handel-C, Oxford University 2001: Mitrion-C, Mitrionics 2003: Bluespec, MIT 2003: MaxJ, Maxeler Technologies 2003: Impulse-C, Impulse Accelerated
    [Show full text]
  • Excel Excel Excel Excel Excel Excel
    CIVL 1112 Origins of Spreadsheets 1/2 Excel Excel Spreadsheets on computers Spreadsheets on computers The word "spreadsheet" came from "spread" in its While other have made contributions to computer-based sense of a newspaper or magazine item that covers spreadsheets, most agree the modern electronic two facing pages, extending across the center fold and spreadsheet was first developed by: treating the two pages as one large one. The compound word "spread-sheet" came to mean the format used to present book-keeping ledgers—with columns for categories of expenditures across the top, invoices listed down the left margin, and the amount of each payment in the cell where its row and column intersect. Dan Bricklin Bob Frankston Excel Excel Spreadsheets on computers Spreadsheets on computers Because of Dan Bricklin and Bob Frankston's Bricklin has spoken of watching his university professor implementation of VisiCalc on the Apple II in 1979 and the create a table of calculation results on a blackboard. IBM PC in 1981, the spreadsheet concept became widely known in the late 1970s and early 1980s. When the professor found an error, he had to tediously erase and rewrite a number of sequential entries in the PC World magazine called VisiCalc the first electronic table, triggering Bricklin to think that he could replicate the spreadsheet. process on a computer, using the blackboard as the model to view results of underlying formulas. His idea became VisiCalc, the first application that turned the personal computer from a hobby for computer enthusiasts into a business tool. Excel Excel Spreadsheets on computers Spreadsheets on computers VisiCalc was the first spreadsheet that combined all VisiCalc went on to become the first killer app, an essential features of modern spreadsheet applications application that was so compelling, people would buy a particular computer just to use it.
    [Show full text]
  • The Computational Attitude in Music Theory
    The Computational Attitude in Music Theory Eamonn Bell Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in the Graduate School of Arts and Sciences COLUMBIA UNIVERSITY 2019 © 2019 Eamonn Bell All rights reserved ABSTRACT The Computational Attitude in Music Theory Eamonn Bell Music studies’s turn to computation during the twentieth century has engendered particular habits of thought about music, habits that remain in operation long after the music scholar has stepped away from the computer. The computational attitude is a way of thinking about music that is learned at the computer but can be applied away from it. It may be manifest in actual computer use, or in invocations of computationalism, a theory of mind whose influence on twentieth-century music theory is palpable. It may also be manifest in more informal discussions about music, which make liberal use of computational metaphors. In Chapter 1, I describe this attitude, the stakes for considering the computer as one of its instruments, and the kinds of historical sources and methodologies we might draw on to chart its ascendance. The remainder of this dissertation considers distinct and varied cases from the mid-twentieth century in which computers or computationalist musical ideas were used to pursue new musical objects, to quantify and classify musical scores as data, and to instantiate a generally music-structuralist mode of analysis. I present an account of the decades-long effort to prepare an exhaustive and accurate catalog of the all-interval twelve-tone series (Chapter 2). This problem was first posed in the 1920s but was not solved until 1959, when the composer Hanns Jelinek collaborated with the computer engineer Heinz Zemanek to jointly develop and run a computer program.
    [Show full text]
  • 1. with Examples of Different Programming Languages Show How Programming Languages Are Organized Along the Given Rubrics: I
    AGBOOLA ABIOLA CSC302 17/SCI01/007 COMPUTER SCIENCE ASSIGNMENT ​ 1. With examples of different programming languages show how programming languages are organized along the given rubrics: i. Unstructured, structured, modular, object oriented, aspect oriented, activity oriented and event oriented programming requirement. ii. Based on domain requirements. iii. Based on requirements i and ii above. 2. Give brief preview of the evolution of programming languages in a chronological order. 3. Vividly distinguish between modular programming paradigm and object oriented programming paradigm. Answer 1i). UNSTRUCTURED LANGUAGE DEVELOPER DATE Assembly Language 1949 FORTRAN John Backus 1957 COBOL CODASYL, ANSI, ISO 1959 JOSS Cliff Shaw, RAND 1963 BASIC John G. Kemeny, Thomas E. Kurtz 1964 TELCOMP BBN 1965 MUMPS Neil Pappalardo 1966 FOCAL Richard Merrill, DEC 1968 STRUCTURED LANGUAGE DEVELOPER DATE ALGOL 58 Friedrich L. Bauer, and co. 1958 ALGOL 60 Backus, Bauer and co. 1960 ABC CWI 1980 Ada United States Department of Defence 1980 Accent R NIS 1980 Action! Optimized Systems Software 1983 Alef Phil Winterbottom 1992 DASL Sun Micro-systems Laboratories 1999-2003 MODULAR LANGUAGE DEVELOPER DATE ALGOL W Niklaus Wirth, Tony Hoare 1966 APL Larry Breed, Dick Lathwell and co. 1966 ALGOL 68 A. Van Wijngaarden and co. 1968 AMOS BASIC FranÇois Lionet anConstantin Stiropoulos 1990 Alice ML Saarland University 2000 Agda Ulf Norell;Catarina coquand(1.0) 2007 Arc Paul Graham, Robert Morris and co. 2008 Bosque Mark Marron 2019 OBJECT-ORIENTED LANGUAGE DEVELOPER DATE C* Thinking Machine 1987 Actor Charles Duff 1988 Aldor Thomas J. Watson Research Center 1990 Amiga E Wouter van Oortmerssen 1993 Action Script Macromedia 1998 BeanShell JCP 1999 AngelScript Andreas Jönsson 2003 Boo Rodrigo B.
    [Show full text]
  • Pioneers of Computing
    Pioneers of Computing В 1980 IEEE Computer Society учредило Золотую медаль (бронзовую) «Вычислительный Пионер» Пионерами учредителями стали 32 члена IEEE Computer Society, связанных с работами по информатике и вычислительным наукам. 1 Pioneers of Computing 1.Howard H. Aiken (Havard Mark I) 2.John V. Atanasoff 3.Charles Babbage (Analytical Engine) 4.John Backus 5.Gordon Bell (Digital) 6.Vannevar Bush 7.Edsger W. Dijkstra 8.John Presper Eckert 9.Douglas C. Engelbart 10.Andrei P. Ershov (theroretical programming) 11.Tommy Flowers (Colossus engineer) 12.Robert W. Floyd 13.Kurt Gödel 14.William R. Hewlett 15.Herman Hollerith 16.Grace M. Hopper 17.Tom Kilburn (Manchester) 2 Pioneers of Computing 1. Donald E. Knuth (TeX) 2. Sergei A. Lebedev 3. Augusta Ada Lovelace 4. Aleksey A.Lyapunov 5. Benoit Mandelbrot 6. John W. Mauchly 7. David Packard 8. Blaise Pascal 9. P. Georg and Edvard Scheutz (Difference Engine, Sweden) 10. C. E. Shannon (information theory) 11. George R. Stibitz 12. Alan M. Turing (Colossus and code-breaking) 13. John von Neumann 14. Maurice V. Wilkes (EDSAC) 15. J.H. Wilkinson (numerical analysis) 16. Freddie C. Williams 17. Niklaus Wirth 18. Stephen Wolfram (Mathematica) 19. Konrad Zuse 3 Pioneers of Computing - 2 Howard H. Aiken (Havard Mark I) – США Создатель первой ЭВМ – 1943 г. Gene M. Amdahl (IBM360 computer architecture, including pipelining, instruction look-ahead, and cache memory) – США (1964 г.) Идеология майнфреймов – система массовой обработки данных John W. Backus (Fortran) – первый язык высокого уровня – 1956 г. 4 Pioneers of Computing - 3 Robert S. Barton For his outstanding contributions in basing the design of computing systems on the hierarchical nature of programs and their data.
    [Show full text]
  • Publications Core Magazine, 2007 Read
    CA PUBLICATIONo OF THE COMPUTERre HISTORY MUSEUM ⁄⁄ SPRINg–SUMMER 2007 REMARKABLE PEOPLE R E scuE d TREAsuREs A collection saved by SAP Focus on E x TRAORdinARy i MAGEs Computers through the Robert Noyce lens of Mark Richards PUBLISHER & Ed I t o R - I n - c hie f THE BEST WAY Karen M. Tucker E X E c U t I V E E d I t o R TO SEE THE FUTURE Leonard J. Shustek M A n A GI n G E d I t o R OF COMPUTING IS Robert S. Stetson A S S o c IA t E E d I t o R TO BROWSE ITS PAST. Kirsten Tashev t E c H n I c A L E d I t o R Dag Spicer E d I t o R Laurie Putnam c o n t RIBU t o RS Leslie Berlin Chris garcia Paula Jabloner Luanne Johnson Len Shustek Dag Spicer Kirsten Tashev d E S IG n Kerry Conboy P R o d U c t I o n ma n ager Robert S. Stetson W E BSI t E M A n AGER Bob Sanguedolce W E BSI t E d ESIG n The computer. In all of human history, rarely has one invention done Dana Chrisler so much to change the world in such a short time. Ton Luong The Computer History Museum is home to the world’s largest collection computerhistory.org/core of computing artifacts and offers a variety of exhibits, programs, and © 2007 Computer History Museum.
    [Show full text]
  • Creativity in Computer Science. in J
    Creativity in Computer Science Daniel Saunders and Paul Thagard University of Waterloo Saunders, D., & Thagard, P. (forthcoming). Creativity in computer science. In J. C. Kaufman & J. Baer (Eds.), Creativity across domains: Faces of the muse. Mahwah, NJ: Lawrence Erlbaum Associates. 1. Introduction Computer science only became established as a field in the 1950s, growing out of theoretical and practical research begun in the previous two decades. The field has exhibited immense creativity, ranging from innovative hardware such as the early mainframes to software breakthroughs such as programming languages and the Internet. Martin Gardner worried that "it would be a sad day if human beings, adjusting to the Computer Revolution, became so intellectually lazy that they lost their power of creative thinking" (Gardner, 1978, p. vi-viii). On the contrary, computers and the theory of computation have provided great opportunities for creative work. This chapter examines several key aspects of creativity in computer science, beginning with the question of how problems arise in computer science. We then discuss the use of analogies in solving key problems in the history of computer science. Our discussion in these sections is based on historical examples, but the following sections discuss the nature of creativity using information from a contemporary source, a set of interviews with practicing computer scientists collected by the Association of Computing Machinery’s on-line student magazine, Crossroads. We then provide a general comparison of creativity in computer science and in the natural sciences. 2. Nature and Origins of Problems in Computer Science December 21, 2004 Computer science is closely related to both mathematics and engineering.
    [Show full text]
  • Arxiv:2106.11534V1 [Cs.DL] 22 Jun 2021 2 Nanjing University of Science and Technology, Nanjing, China 3 University of Southampton, Southampton, U.K
    Noname manuscript No. (will be inserted by the editor) Turing Award elites revisited: patterns of productivity, collaboration, authorship and impact Yinyu Jin1 · Sha Yuan1∗ · Zhou Shao2, 4 · Wendy Hall3 · Jie Tang4 Received: date / Accepted: date Abstract The Turing Award is recognized as the most influential and presti- gious award in the field of computer science(CS). With the rise of the science of science (SciSci), a large amount of bibliographic data has been analyzed in an attempt to understand the hidden mechanism of scientific evolution. These include the analysis of the Nobel Prize, including physics, chemistry, medicine, etc. In this article, we extract and analyze the data of 72 Turing Award lau- reates from the complete bibliographic data, fill the gap in the lack of Turing Award analysis, and discover the development characteristics of computer sci- ence as an independent discipline. First, we show most Turing Award laureates have long-term and high-quality educational backgrounds, and more than 61% of them have a degree in mathematics, which indicates that mathematics has played a significant role in the development of computer science. Secondly, the data shows that not all scholars have high productivity and high h-index; that is, the number of publications and h-index is not the leading indicator for evaluating the Turing Award. Third, the average age of awardees has increased from 40 to around 70 in recent years. This may be because new breakthroughs take longer, and some new technologies need time to prove their influence. Besides, we have also found that in the past ten years, international collabo- ration has experienced explosive growth, showing a new paradigm in the form of collaboration.
    [Show full text]