Computers Are Networked

Total Page:16

File Type:pdf, Size:1020Kb

Computers Are Networked Chapter 11 Computers are networked In the very early days, according to a remark often attributed to Howard Aiken, some of the people in business believed that “only a very small number of computers would be needed to serve the needs of the whole world, per- haps a dozen, with eight or ten for the United States.”. Although it is somewhat misquoted, the fact is that there used to be very few computers, and they were quite expensive. As a result, it was not very accessible to the general public. 1 It certainly has changed.... This situation started to change in the early 1980’s, when IBM introduced its personal com- puter (IBM PC) for use in the home, office and schools. With an introduction of $1,995, followed by many clones, computers suddenly became af- fordable. The number of personal computers in use was more than doubled from 2 million in 1981 to 5.5 million in 1982. Ten years later, 65 million PCs were being used. The computers were getting smaller: from desk- top to laptop, notebook, to tablets today, while retaining its processing power, and the price keeps on dropping. We are even talking about $100 notebook for the kids. Thus, many of us can now afford to have a computer. 2 A big thing All such computers, big or small, were con- nected together into networks to share hard- ware such as a printer, software such as Alice, and, more importantly, information. Just look at any cluster on campus: twenty plus computers share one or two printers, they share some of the applications, such as the Microsoft Office suite, as well as the data: In- deed, when any one tries to look at your grade, they are looking at the same document. Eventually, a global web of computer, the In- ternet, links computers worldwide into a single network of information. 3 How did we get Internet? In October 1962, Joseph Licklider was appointed head of the United States Department of De- fense’s Advanced Research Projects Agency (DARPA), within the Information Processing Office. Joseph formed an informal group within DARPA to further computer research, including installing three network terminals: one for System De- velopment Corporation in Santa Monica, one for Project Genie at the UC Berkeley and one for the Compatible Time-Sharing System project at the MIT. Question: What is Timing sharing? Answers: When you have to do multiple things, they take turns. 4 Then what? Several research programs began to explore and articulate principles of networking between physically separate networks, leading to the de- velopment of the packet switching model of digital networking. The basic idea of packet switching is this: When sending data between two computers connected to a network, we will cut the data into la- beled packets, select paths in between, often the shortest ones, send the packets over to the destination, possibly along different paths, then reassemble them back to the original doc- ument. 5 Finally,... Such research led to the development of sev- eral packet-switching networks in the late 1960s and 1970s, including the ARPANET. But, these networks were still disjointed sepa- rate networks, served only by limited connec- tion (gateways) between networks. Further efforts were then made to develop a protocol(a set of rules) for internetworking, where multiple different networks could be joined together into a bigger network. This spread of internetworking began to form a global network, i.e., Internet, based on stan- dardized protocols officially implemented in 1982. 6 Local area network At the same time, the technology of inter- connecting many networks of a much smaller scale(local area network) enabled the sharing of software, hardware and information in a lo- cal area such as Memorial 213. Thus, as we discussed in the Data structure module, Internet is really a complicated graph structure that interconnect computers, many form their own local area networks. 7 Here comes the Web... As the Internet grew through the 1980s and early 1990s, people felt the need to organize files and information spread in such a incre- mentally complicated structure so that they can be easily found. Projects such as Gopher, WAIS, and the FTP Archive list made early attempts to create ways to organize distributed data, but, unfortunately, they all failed to accommodate all the existing data types. Furthermore, none was user friendly, i.e., they were pretty difficult to use. 8 A better interface One of the most promising user interface paradigms during this period was hypertext, which we use today in web pages, i.e., when you click a link, it goes to somewhere in the Net. It consists of two pieces, one is the name of this link, which the user knows; and the other is the where- abouts of this link, which is technical and the user does not need to know. This technology actually occurred earlier, Go- pher implemented the first commonly-used hy- pertext interface to the Internet. 9 Just a bit technical When you launch such a browser, you are con- nected to its Home page, the start point of your navigation. When you click at any of its hypertext links, with its hidden address in the form of a URL, the browser will display that web page, which again contains many links, and the navigation continues. A typical URL is the following: http://www.nps.gov/yose/planyourvisit/rvcamping.htm Here, the http stands for hypertext transfer protocol, which tells the browser how to talk to that particular machine. The site www.nps.gov tells who owns this web site, and implicitly where the machine is, via an address book. The rest of the address yose/planyourvisit/rvcamping.htm provides the name of the web page. 10 ... and the other stuff In 1989, while working at CERN, Tim Berners- Lee invented a network-based implementation of the hypertext concept, which led to the eventual development of the World Wide Web. A turning point is the Mosaic graphic web browser that a team based in University of Illinois at Urbana-Champaign, developed in 1993, led by Marc Andreessen, and completed by the Netscape company. Yahoo was founded in 1995, Google started in 1996, all doing business to help us to find what we want in the Web. Then, the software giant Microsoft introduced IE in August 2005. The rest is history. 11 Other Internet service 1. FTP(File Transfer Protocol): This is for you to connect to a remote computer, an FTP site, to transfer files to your own computers. 2. Telnet: This is for you to connect to a remote computer. For example, when you are home, but want to log on to a mainframe com- puter such as oz or turing. It is especially useful if you keep your files there, or to get informa- tion from a large data depository, such as a library. If you want to log on to your M: drive from a remote site, and transfer files in between, you need to download a piece of software, e.g., PuTTY. 3. Newsgroups: This is for you to join or just read electronic news bulletins. 4. Chat rooms: This is for you to join a live discussion about anything. 12 Internet Service Provider (ISP) An ISP is a company that provides the ser- vice so that you can connect to the Internet, just like what a phone company does when you want to talk to other people. Examples include AOL, AT&T, MSN, and many of your local providers, such as Metrocast. The process is very similar to order other com- munication service. Once they come to install the equipment, and hook you up, you log on and in. 13 Spam: unwanted email These days, we often have to spend lots of time in discovering, and deleting email, that we don’t want. It is estimated that over 60 % of the emails are spam mails. In fact, for the California-based Postini Inc., about 85% percent of the roughly 1 billion pieces of e-mail the company handles each week are spams. AOL blocks about 2.4 billion pieces of spam mail every day. It costs about 10 billion dollars to fight spam mails. 14 Fight Spam Besides the measures as suggested in pp. 53, such as deletion w/o opening, never reply to such a message, and add on a filter, etc., Some legal actions have also been taken. President Bush signed into law which makes span illegal. Some states, such as California, already has strict anti-spam laws. But, as we saw, the hard-core spammers have yet to show any sign to clean their act, which now pays about $200 to $500 for sending out a million spam messages. 15 Black Penny project Some serious efforts, e.g., the Penny Black Project, have been made to destroy the busi- ness model of email spam. The essential idea is to let the sender spend more time, but not the receiver. “If I don’t know you, you have to prove to me that you spent ten seconds of your CPU time just for me, and just for this message.” In practice, an algorithm is put in every email system so that it takes several seconds to send it out to any receiver, which will make the practice of spamming essentially infeasible. Microsoft, among others, has been working on such a project. 16 Activities Activities Write a one-page single-space es- say either for, or against, one of the following statements: 1. In this information age, English will become a universal language.
Recommended publications
  • A Postmortem for a Time Sharing System
    A POSTMORTEM FOR A TIME SHARING SYSTEM HOWARD EWING STURGIS CSL 74-1 JANUARY, 1974 Abstract This thesis describes a time sharing system constructed by a project at the University of California, Berkeley Campus, Computer Center. The project was of modest size, consuming about 30 man years. The resulting system was used by a number of programmers. The system was designed for a commercially available computer, the Control Data 6400 with extended core store. The system design was based on several fundamental ideas, including: specification of the entire system as an abstract machine, a capability based protection system, mapped address space, and layered implementation. The abstract machine defined by the first implementation layer provided 8 types of abstractly defined objects and about 100 actions to manipulate them. Subsequent layers provided a few very complicated additional types. Many of the fundamental ideas served us well, particularly the concept that the system defines an abstract machine, and capability based protection. However, the attempt to provide a mapped address space using unsuitable hardware was a disaster. This thesis includes software and hardware proposals to increase the efficiency of representing an abstract machine and providing capability based protection. Also included is a description of a crash recovery consistency prob- lem for files which reside in several levels of storage, together with a solution that we used. XEROX PALO ALTO RESEARCH CENTER 3180 PORTER DRIVE/PALO ALTO/CALIFORNIA 94304 ACKNOWLEDGEMENTS∗ First, I thank Professor James Morris, my dissertation committee chairman, for many hours of discussions and painstaking reading of many drafts. Second, I thank the other members of my dissertation committee, Professor R.
    [Show full text]
  • The Evolution of the Unix Time-Sharing System*
    The Evolution of the Unix Time-sharing System* Dennis M. Ritchie Bell Laboratories, Murray Hill, NJ, 07974 ABSTRACT This paper presents a brief history of the early development of the Unix operating system. It concentrates on the evolution of the file system, the process-control mechanism, and the idea of pipelined commands. Some attention is paid to social conditions during the development of the system. NOTE: *This paper was first presented at the Language Design and Programming Methodology conference at Sydney, Australia, September 1979. The conference proceedings were published as Lecture Notes in Computer Science #79: Language Design and Programming Methodology, Springer-Verlag, 1980. This rendition is based on a reprinted version appearing in AT&T Bell Laboratories Technical Journal 63 No. 6 Part 2, October 1984, pp. 1577-93. Introduction During the past few years, the Unix operating system has come into wide use, so wide that its very name has become a trademark of Bell Laboratories. Its important characteristics have become known to many people. It has suffered much rewriting and tinkering since the first publication describing it in 1974 [1], but few fundamental changes. However, Unix was born in 1969 not 1974, and the account of its development makes a little-known and perhaps instructive story. This paper presents a technical and social history of the evolution of the system. Origins For computer science at Bell Laboratories, the period 1968-1969 was somewhat unsettled. The main reason for this was the slow, though clearly inevitable, withdrawal of the Labs from the Multics project. To the Labs computing community as a whole, the problem was the increasing obviousness of the failure of Multics to deliver promptly any sort of usable system, let alone the panacea envisioned earlier.
    [Show full text]
  • The People Who Invented the Internet Source: Wikipedia's History of the Internet
    The People Who Invented the Internet Source: Wikipedia's History of the Internet PDF generated using the open source mwlib toolkit. See http://code.pediapress.com/ for more information. PDF generated at: Sat, 22 Sep 2012 02:49:54 UTC Contents Articles History of the Internet 1 Barry Appelman 26 Paul Baran 28 Vint Cerf 33 Danny Cohen (engineer) 41 David D. Clark 44 Steve Crocker 45 Donald Davies 47 Douglas Engelbart 49 Charles M. Herzfeld 56 Internet Engineering Task Force 58 Bob Kahn 61 Peter T. Kirstein 65 Leonard Kleinrock 66 John Klensin 70 J. C. R. Licklider 71 Jon Postel 77 Louis Pouzin 80 Lawrence Roberts (scientist) 81 John Romkey 84 Ivan Sutherland 85 Robert Taylor (computer scientist) 89 Ray Tomlinson 92 Oleg Vishnepolsky 94 Phil Zimmermann 96 References Article Sources and Contributors 99 Image Sources, Licenses and Contributors 102 Article Licenses License 103 History of the Internet 1 History of the Internet The history of the Internet began with the development of electronic computers in the 1950s. This began with point-to-point communication between mainframe computers and terminals, expanded to point-to-point connections between computers and then early research into packet switching. Packet switched networks such as ARPANET, Mark I at NPL in the UK, CYCLADES, Merit Network, Tymnet, and Telenet, were developed in the late 1960s and early 1970s using a variety of protocols. The ARPANET in particular led to the development of protocols for internetworking, where multiple separate networks could be joined together into a network of networks. In 1982 the Internet Protocol Suite (TCP/IP) was standardized and the concept of a world-wide network of fully interconnected TCP/IP networks called the Internet was introduced.
    [Show full text]
  • ARPANET and Its Boundary Devices: Modems, Imps, and the Inter-Structuralism of Infrastructures
    Internet Histories Digital Technology, Culture and Society ISSN: 2470-1475 (Print) 2470-1483 (Online) Journal homepage: http://www.tandfonline.com/loi/rint20 ARPANET and its boundary devices: modems, IMPs, and the inter-structuralism of infrastructures Fenwick McKelvey & Kevin Driscoll To cite this article: Fenwick McKelvey & Kevin Driscoll (2018): ARPANET and its boundary devices: modems, IMPs, and the inter-structuralism of infrastructures, Internet Histories, DOI: 10.1080/24701475.2018.1548138 To link to this article: https://doi.org/10.1080/24701475.2018.1548138 Published online: 28 Dec 2018. Submit your article to this journal Article views: 2 View Crossmark data Full Terms & Conditions of access and use can be found at http://www.tandfonline.com/action/journalInformation?journalCode=rint20 INTERNET HISTORIES https://doi.org/10.1080/24701475.2018.1548138 ARPANET and its boundary devices: modems, IMPs, and the inter-structuralism of infrastructures Fenwick McKelveya and Kevin Driscollb aDepartment of Communication Studies, Concordia University, Montreal, Quebec, Canada; bMedia Studies, University of Virginia, Charlottesville, Virginia, USA ABSTRACT ARTICLE HISTORY Our paper focuses on the Interface Message Processor (IMP), an Received 21 April 2018 important device in the history of ARPANET. Designed as the Revised 1 November 2018 interface between ARPANET nodes and the common carrier tele- Accepted 2 November 2018 phone system, the IMP actualized the ARPANET as an experimen- KEYWORDS tal packet-switching communication system. We conceptualize ’ ARPANET; infrastructure; the IMP as historical boundary object that exposes ARPANET s boundary objects; gateways; close relationship to the telephone system. Our analysis offers a internetworking novel history of ARPANET as a repurposing of the existing tele- phone infrastructure.
    [Show full text]
  • Notes on the History of Fork-And-Join Linus Nyman and Mikael Laakso Hanken School of Economics
    Notes on the history of fork-and-join Linus Nyman and Mikael Laakso Hanken School of Economics As part of a PhD on code forking in open source software, Linus Nyman looked into the origins of how the practice came to be called forking.1 This search led back to the early history of the fork system call. Having not previously seen such a history published, this anecdote looks back at the birth of the fork system call to share what was learned, as remembered by its pioneers. The fork call allows a process (or running program) to create new processes. The original is deemed the parent process and the newly created one its child. On multiprocessor systems, these processes can run concurrently in parallel.2 Since its birth 50 years ago, the fork has remained a central element of modern computing, both with regards to software development principles and, by extension, to hardware design, which increasingly accommodates parallelism in process execution. The fork system call is imagined The year was 1962. Melvin Conway, later to become known for “Conway’s Law,”3 was troubled by what seemed an unnecessary inefficiency in computing. As Conway recalls:4 I was in the US Air Force at the time – involved with computer procurement – and I noticed that no vendor was distinguishing between “processor” and “process.” That is, when a proposal involved a multiprocessor design, each processor was committed to a single process. By late 1962, Conway had begun contemplating the idea of using a record to carry the status of a process.
    [Show full text]
  • ADVANCED INTELLECT-AUGMENTATION TECHNIQUES by D
    - -, 3 i.? NASACONTRACTOR NASA REPORT C.1 LOAN COPY RETURN TO AFWL (DOWL) KIRTUND AFB, N. M, ADVANCED INTELLECT-AUGMENTATION TECHNIQUES by D. C. Engelburt und Staff of Augmegztution Reseurch Center Prepared by STANFORD RESEARCH INSTITUTE Menlo Park, Calif. 94025 .for LangZey Research Center - .~ - .. - - . - - "_ - -- 1. ReportNo. I 2. Government Accession No. 73.Recipient's Catalog No. NASA CR-1827 """ ~~~ ~~ 1 .~ ~~~ I 4. Title and Subtitle 5. Report Date February 1972 ADVANCED INTELLECT-AUGMENTATION TECHNIQUES 6. Performing Organization Code __~___. .___ I 7. Author(s) 8. Performing Organization Report No. D. C. Engelbart and Staff of Augmentation Research Center SRI Project 7079 . -~ . - ~ 10. Work Unit No. 9. Performing Organization Name and Address Stanford Research Institute 11. Contract orGrant No. MenloPark, California 94025 I NAS1-7897 13. Type of Report and Period Covered -~ ~ ~ 12. Sponsoring Agency Name and Address Contractor Report National Aeronautics and Space Administration 14. Sponsoring Agency Code Washington, D.C.20546 __~__-~ 15. Supplementary Notes This report covers the extension of research previously reported in NASA CR-1270. ~ "- 16. Abstract This report covers a two-year project, at the eleventh year of a growing, multiproject program that is exploring the value of computer aids to augmenting human intellectual capability. Outlined briefly are the background and the "bootstrapping" nature of the program, its resources, ai the activities it has undertaken in pursuit of its goals. User experience in applying our augmentation tools and techniques to various normal working tasks within our Center is described so as to convey a subjective impression of what it is like to work in an augmented environment.
    [Show full text]
  • The Computer History Simulation Project
    The Computer History Simulation Project The Computer History Simulation Project The Computer History Simulation Project is a loose Internet-based collective of people interested in restoring historically significant computer hardware and software systems by simulation. The goal of the project is to create highly portable system simulators and to publish them as freeware on the Internet, with freely available copies of significant or representative software. Simulators SIMH is a highly portable, multi-system simulator. ● Download the latest sources for SIMH (V3.5-1 updated 15-Oct-2005 - see change log). ● Download a zip file containing Windows executables for all the SIMH simulators. The VAX and PDP-11 are compiled without Ethernet support. Versions with Ethernet support are available here. If you download the executables, you should download the source archive as well, as it contains the documentation and other supporting files. ● If your host system is Alpha/VMS, and you want Ethernet support, you need to download the VMS Pcap library and execlet here. SIMH implements simulators for: ● Data General Nova, Eclipse ● Digital Equipment Corporation PDP-1, PDP-4, PDP-7, PDP-8, PDP-9, PDP-10, PDP-11, PDP- 15, VAX ● GRI Corporation GRI-909 ● IBM 1401, 1620, 1130, System 3 ● Interdata (Perkin-Elmer) 16b and 32b systems ● Hewlett-Packard 2116, 2100, 21MX ● Honeywell H316/H516 ● MITS Altair 8800, with both 8080 and Z80 ● Royal-Mcbee LGP-30, LGP-21 ● Scientific Data Systems SDS 940 Also available is a collection of tools for manipulating simulator file formats and for cross- assembling code for the PDP-1, PDP-7, PDP-8, and PDP-11.
    [Show full text]
  • A History of Silicon Valley the Greatest Creation of Wealth in History (An Immoral Tale) Being a Presentation by Piero Scaruffi
    A History of Silicon Valley The Greatest Creation of Wealth in History (An immoral tale) being a presentation by piero scaruffi www.scaruffi.com adapted from the book “A History of Silicon Valley” Piero Scaruffi • Cultural Historian • Cognitive Scientist • Blogger • Poet • www.scaruffi.com www.scaruffi.com 2 This is Part 2 • See http://www.scaruffi.com/svhistory for the index of this Powerpoint presentation and links to the other parts – 1900-1960 – The 1960s – The 1970s – The 1980s – The 1990s – The 2000s www.scaruffi.com 3 What the book is about… • The book is a history of the high-tech industry in the San Francisco Bay Area (of which Silicon Valley is currently the most famous component) www.scaruffi.com 4 Semiconductors – Fairchild Semiconductors’ planar integrated circuit (1961) – Fairchild Semiconductors employees: Don Farina, Don Valentine, Charles Sporck, Jerry Sanders, Jack Gifford, Mike Markkula – Signetics (1961), first Fairchild spinoff – Main customers of integrated circuits: the Air Force and NASA www.scaruffi.com 5 Life Sciences • Stanford hires Carl Djerassi (1959), inventor of the birth-control pill • Alejandro Zaffaroni’s Syntex relocates to the Stanford Industrial Park (1963) www.scaruffi.com 6 Meanwhile elsewhere… • New York – IBM 7000 transistorized series (1960) – IBM’s SABRE (1960), the first online transaction processing, an adaptation of SAGE to automating American Airlines' reservation system – GE’s IDS (1961), the first database management system • Boston (MIT) – CTSS (1961), the first time-sharing system – "Spacewar" (1962), the first computer game – Ivan Sutherland’s "Sketchpad“ (1963), the first computer program with a GUI www.scaruffi.com 7 Meanwhile elsewhere… • US Government – Paul Baran (Rand Corp): a distributed network of computers can survive a nuclear strike (1962) – Ted Nelson (Harvard Univ): hypertext (1965) – Joseph Licklider (DARPA’s IPTO) funds Project MAC for A.I.
    [Show full text]
  • Unix: Building a Development Environment from Scratch
    Unix: Building a Development Environment from Scratch Warren Toomey Abstract In April 1969, as part of AT&T’s withdrawal from the Multics project, the researchers involved had their “pleasant” development environment taken from them. Bereft of their “toy”, the ex-Multics researchers began to cast about for a re- placement. Having found nothing suitable, Ken Thompson chose to write one from scratch. By the middle of 1969, he had created a self-hosting operating system on a discarded PDP-7 minicomputer. This was Unix, an operating system whose legacy remains with us today. This paper looks at the creation of Unix after AT&T’s depar- ture from the Multics project, the features and innovations in the PDP-7 version of Unix, and the work done in 2016 to restore a working version of PDP-7 Unix from the available source code. 1 The Motivation behind Unix’s Development 1969 was not a good year for the Bell Labs researchers at AT&T who were involved in the Multics project. Multics was an attempt to refine and extend many of the contemporary ideas in operating systems (e.g. virtual memory, multitasking) and to build an operating system utility with “a view of continuous operation analogous to that of the electric power and telephone companies” [3]. The Multics designers had set an ambitious list of features for the system, includ- ing: • Virtual, segmented memory, so that a process could access more memory than physically available on the system • File-mapped memory, for persistent storage of in-memory data structures • A hierachical filesystem with
    [Show full text]
  • The UNIX Time-Sharing System*
    The UNIX Time-Sharing System* D. M. Ritchie and K. Thompson ABSTRACT UNIX² is a general-purpose, multi-user, interactive operating system for the larger Digital Equipment Corporation PDP-11 and the Interdata 8/32 computers. It offers a number of features seldom found even in larger operating systems, including i A hierarchical ®le system incorporating demountable volumes, ii Compatible ®le, device, and inter-process I/O, iii The ability to initiate asynchronous processes, iv System command language selectable on a per-user basis, v Over 100 subsystems including a dozen languages, vi High degree of portability. This paper discusses the nature and implementation of the ®le system and of the user command interface. 1. INTRODUCTION There have been four versions of the UNIX time-sharing system. The earliest (circa 1969-70) ran on the Digital Equipment Corporation PDP-7 and -9 computers. The second version ran on the unpro- tected PDP-11/20 computer. The third incorporated multiprogramming and ran on the PDP-11/34, /40, /45, /60, and /70 computers; it is the one described in the previously published version of this paper, and is also the most widely used today. This paper describes only the fourth, current system that runs on the PDP-11/70 and the Interdata 8/32 computers. In fact, the differences among the various systems is rather small; most of the revisions made to the originally published version of this paper, aside from those con- cerned with style, had to do with details of the implementation of the ®le system. Since PDP-11 UNIX became operational in February, 1971, over 600 installations have been put into service.
    [Show full text]
  • Compatible Time-Sharing System (1961-1973) Fiftieth Anniversary
    Compatible Time-Sharing System (1961-1973) Fiftieth Anniversary Commemorative Overview The Compatible Time Sharing System (1961–1973) Fiftieth Anniversary Commemorative Overview The design of the cover, half-title page (reverse side of this page), and main title page mimics the design of the 1963 book The Compatible Time-Sharing System: A Programmer’s Guide from the MIT Press. The Compatible Time Sharing System (1961–1973) Fiftieth Anniversary Commemorative Overview Edited by David Walden and Tom Van Vleck With contributions by Fernando Corbató Marjorie Daggett Robert Daley Peter Denning David Alan Grier Richard Mills Roger Roach Allan Scherr Copyright © 2011 David Walden and Tom Van Vleck All rights reserved. Single copies may be printed for personal use. Write to us at [email protected] for a PDF suitable for two-sided printing of multiple copies for non-profit academic or scholarly use. IEEE Computer Society 2001 L Street N.W., Suite 700 Washington, DC 20036-4928 USA First print edition June 2011 (web revision 03) The web edition at http://www.computer.org/portal/web/volunteercenter/history is being updated from the print edition to correct mistakes and add new information. The change history is on page 50. To Fernando Corbató and his MIT collaborators in creating CTSS Contents Preface . ix Acknowledgments . ix MIT nomenclature and abbreviations . x 1 Early history of CTSS . 1 2 The IBM 7094 and CTSS at MIT . 5 3 Uses . 17 4 Memories and views of CTSS . 21 Fernando Corbató . 21 Marjorie Daggett . 22 Robert Daley . 24 Richard Mills . 26 Tom Van Vleck . 31 Roger Roach .
    [Show full text]
  • The Evolution of the Unix Time-Sharing System*
    The Evolution of the Unix Time-sharing System* Dennis M. Ritchie Bell Laboratories Murray Hill, New Jersey 07974 ABSTRACT This paper presents a brief history of the early development of the Unix operating system. It concentrates on the evolution of the file system, the process-control mechanism, and the idea of pipelined commands. Some atten- tion is paid to social conditions during the development of the system. Introduction During the past few years, the Unix operating system has come into wide use, so wide that its very name has become a trademark of Bell Laboratories. Its important characteristics have become known to many people. It has suffered much rewriting and tinkering since the first pub- lication describing it in 1974 [1], but few fundamental changes. However, Unix was born in 1969 not 1974, and the account of its development makes a little-known and perhaps instructive story. This paper presents a technical and social history of the evolution of the system. Origins For computer science at Bell Laboratories, the period 1968-1969 was somewhat unsettled. The main reason for this was the slow, though clearly inevitable, withdrawal of the Labs from the Multics project. To the Labs computing community as a whole, the problem was the increasing obviousness of the failure of Multics to deliver promptly any sort of usable system, let alone the panacea envisioned earlier. For much of this time, the Murray Hill Computer Center was also running a costly GE 645 machine that inadequately simulated the GE 635. Another shake-up that occurred during this period was the organizational separation of computing services and com- puting research.
    [Show full text]