372.9Kb Application/Pdf
Total Page:16
File Type:pdf, Size:1020Kb

Load more
Recommended publications
-
Proof-Carrying Authentication∗
Proof-Carrying Authentication∗ Andrew W. Appel and Edward W. Felten Secure Internet Programming Laboratory Department of Computer Science Princeton University Princeton, NJ 08544 USA August 9, 1999 Abstract Statements in these frameworks are represented as data structures that are often digitally signed We have designed and implemented a general to ensure their integrity. and powerful distributed authentication frame- work based on higher-order logic. Authenti- Several authentication frameworks exist; we cation frameworks — including Taos, SPKI, mention a few here as examples. The Taos SDSI, and X.509 — have been explained using operating system provided support for secure logic. We show that by starting with the logic, remote procedure call and data structures to we can implement these frameworks, all in the represent authority and identity [6]. X.509 [15] same concise and efficient system. Because our is a widely-used standard for expressing and us- logic has no decision procedure — although ing digital certificates. SPKI [4] and SDSI [14] proof checking is simple — users of the frame- (since merged under the joint name SPKI) work must submit proofs with their requests. were reactions to the perceived complexity of X.509; in both cases the ‘S’ stands for ‘simple.’ 1 Introduction PolicyMaker [3] is a language for expressing se- curity policies; it can be applied to distributed Distributed authentication frameworks allow security policies. Kerberos [12], unlike the other sharing of access to resources across adminis- frameworks, uses symmetric-key encryption to trative boundaries in a distributed system. The authenticate users. Each framework has a main abstractions they support are name-to- differerent semantics and offers a different kind key bindings, access control, and delegation. -
Computers Are Networked
Chapter 11 Computers are networked In the very early days, according to a remark often attributed to Howard Aiken, some of the people in business believed that “only a very small number of computers would be needed to serve the needs of the whole world, per- haps a dozen, with eight or ten for the United States.”. Although it is somewhat misquoted, the fact is that there used to be very few computers, and they were quite expensive. As a result, it was not very accessible to the general public. 1 It certainly has changed.... This situation started to change in the early 1980’s, when IBM introduced its personal com- puter (IBM PC) for use in the home, office and schools. With an introduction of $1,995, followed by many clones, computers suddenly became af- fordable. The number of personal computers in use was more than doubled from 2 million in 1981 to 5.5 million in 1982. Ten years later, 65 million PCs were being used. The computers were getting smaller: from desk- top to laptop, notebook, to tablets today, while retaining its processing power, and the price keeps on dropping. We are even talking about $100 notebook for the kids. Thus, many of us can now afford to have a computer. 2 A big thing All such computers, big or small, were con- nected together into networks to share hard- ware such as a printer, software such as Alice, and, more importantly, information. Just look at any cluster on campus: twenty plus computers share one or two printers, they share some of the applications, such as the Microsoft Office suite, as well as the data: In- deed, when any one tries to look at your grade, they are looking at the same document. -
A Postmortem for a Time Sharing System
A POSTMORTEM FOR A TIME SHARING SYSTEM HOWARD EWING STURGIS CSL 74-1 JANUARY, 1974 Abstract This thesis describes a time sharing system constructed by a project at the University of California, Berkeley Campus, Computer Center. The project was of modest size, consuming about 30 man years. The resulting system was used by a number of programmers. The system was designed for a commercially available computer, the Control Data 6400 with extended core store. The system design was based on several fundamental ideas, including: specification of the entire system as an abstract machine, a capability based protection system, mapped address space, and layered implementation. The abstract machine defined by the first implementation layer provided 8 types of abstractly defined objects and about 100 actions to manipulate them. Subsequent layers provided a few very complicated additional types. Many of the fundamental ideas served us well, particularly the concept that the system defines an abstract machine, and capability based protection. However, the attempt to provide a mapped address space using unsuitable hardware was a disaster. This thesis includes software and hardware proposals to increase the efficiency of representing an abstract machine and providing capability based protection. Also included is a description of a crash recovery consistency prob- lem for files which reside in several levels of storage, together with a solution that we used. XEROX PALO ALTO RESEARCH CENTER 3180 PORTER DRIVE/PALO ALTO/CALIFORNIA 94304 ACKNOWLEDGEMENTS∗ First, I thank Professor James Morris, my dissertation committee chairman, for many hours of discussions and painstaking reading of many drafts. Second, I thank the other members of my dissertation committee, Professor R. -
A Model of Inheritance for Declarative Visual Programming Languages
An Abstract Of The Dissertation Of Rebecca Djang for the degree of Doctor of Philosophy in Computer Science presented on December 17, 1998. Title: Similarity Inheritance: A Model of Inheritance for Declarative Visual Programming Languages. Abstract approved: Margaret M. Burnett Declarative visual programming languages (VPLs), including spreadsheets, make up a large portion of both research and commercial VPLs. Spreadsheets in particular enjoy a wide audience, including end users. Unfortunately, spreadsheets and most other declarative VPLs still suffer from some of the problems that have been solved in other languages, such as ad-hoc (cut-and-paste) reuse of code which has been remedied in object-oriented languages, for example, through the code-reuse mechanism of inheritance. We believe spreadsheets and other declarative VPLs can benefit from the addition of an inheritance-like mechanism for fine-grained code reuse. This dissertation first examines the opportunities for supporting reuse inherent in declarative VPLs, and then introduces similarity inheritance and describes a prototype of this model in the research spreadsheet language Forms/3. Similarity inheritance is very flexible, allowing multiple granularities of code sharing and even mutual inheritance; it includes explicit representations of inherited code and all sharing relationships, and it subsumes the current spreadsheet mechanisms for formula propagation, providing a gradual migration from simple formula reuse to more sophisticated uses of inheritance among objects. Since the inheritance model separates inheritance from types, we investigate what notion of types is appropriate to support reuse of functions on different types (operation polymorphism). Because it is important to us that immediate feedback, which is characteristic of many VPLs, be preserved, including feedback with respect to type errors, we introduce a model of types suitable for static type inference in the presence of operation polymorphism with similarity inheritance. -
Unifying Theories of Objects
Unifying Theories of Objects Michael Anthony Smith1;2 and Jeremy Gibbons1 1 Oxford University, UK. 2 Systems Assurance Group, QinetiQ, UK. [email protected] [email protected] Abstract. We present an approach to modelling Abadi{Cardelli-style object calculi as Unifying Theories of Programming (UTP) designs. Here we provide a core object calculus with an operational small-step evalua- tion rule semantics, and a corresponding UTP model with a denotational relational predicate semantics. For clarity, the UTP model is defined in terms of an operand stack, which is used to store the results of sub- programs. Models of a less operational nature are briefly discussed. The consistency of the UTP model is demonstrated by a structural induction proof over the operations of the core object calculus. Overall, our UTP model is intended to provide facilities for encoding both object-based and class-based languages. 1 Introduction Hoare and He's Unifying Theories of Programming (UTP) [6] can be used to formally define how results produced in one formal model can be translated as assumptions to another formal model. Essentially, programs are considered to be predicates that relate the values of their observable input and output variables (their alphabet). For example, the increment program x := x + 1 is typically defined by the relational predicate x 0 = x + 1, where predicate variables x and x 0 denote the input and output values of the program variable x. In general, the alphabet of a program P is denoted by αP; it is the disjoint union of P's input and output sets (inαP and outαP respectively), which in the case of the example is the set fx; x 0g. -
The Evolution of the Unix Time-Sharing System*
The Evolution of the Unix Time-sharing System* Dennis M. Ritchie Bell Laboratories, Murray Hill, NJ, 07974 ABSTRACT This paper presents a brief history of the early development of the Unix operating system. It concentrates on the evolution of the file system, the process-control mechanism, and the idea of pipelined commands. Some attention is paid to social conditions during the development of the system. NOTE: *This paper was first presented at the Language Design and Programming Methodology conference at Sydney, Australia, September 1979. The conference proceedings were published as Lecture Notes in Computer Science #79: Language Design and Programming Methodology, Springer-Verlag, 1980. This rendition is based on a reprinted version appearing in AT&T Bell Laboratories Technical Journal 63 No. 6 Part 2, October 1984, pp. 1577-93. Introduction During the past few years, the Unix operating system has come into wide use, so wide that its very name has become a trademark of Bell Laboratories. Its important characteristics have become known to many people. It has suffered much rewriting and tinkering since the first publication describing it in 1974 [1], but few fundamental changes. However, Unix was born in 1969 not 1974, and the account of its development makes a little-known and perhaps instructive story. This paper presents a technical and social history of the evolution of the system. Origins For computer science at Bell Laboratories, the period 1968-1969 was somewhat unsettled. The main reason for this was the slow, though clearly inevitable, withdrawal of the Labs from the Multics project. To the Labs computing community as a whole, the problem was the increasing obviousness of the failure of Multics to deliver promptly any sort of usable system, let alone the panacea envisioned earlier. -
The People Who Invented the Internet Source: Wikipedia's History of the Internet
The People Who Invented the Internet Source: Wikipedia's History of the Internet PDF generated using the open source mwlib toolkit. See http://code.pediapress.com/ for more information. PDF generated at: Sat, 22 Sep 2012 02:49:54 UTC Contents Articles History of the Internet 1 Barry Appelman 26 Paul Baran 28 Vint Cerf 33 Danny Cohen (engineer) 41 David D. Clark 44 Steve Crocker 45 Donald Davies 47 Douglas Engelbart 49 Charles M. Herzfeld 56 Internet Engineering Task Force 58 Bob Kahn 61 Peter T. Kirstein 65 Leonard Kleinrock 66 John Klensin 70 J. C. R. Licklider 71 Jon Postel 77 Louis Pouzin 80 Lawrence Roberts (scientist) 81 John Romkey 84 Ivan Sutherland 85 Robert Taylor (computer scientist) 89 Ray Tomlinson 92 Oleg Vishnepolsky 94 Phil Zimmermann 96 References Article Sources and Contributors 99 Image Sources, Licenses and Contributors 102 Article Licenses License 103 History of the Internet 1 History of the Internet The history of the Internet began with the development of electronic computers in the 1950s. This began with point-to-point communication between mainframe computers and terminals, expanded to point-to-point connections between computers and then early research into packet switching. Packet switched networks such as ARPANET, Mark I at NPL in the UK, CYCLADES, Merit Network, Tymnet, and Telenet, were developed in the late 1960s and early 1970s using a variety of protocols. The ARPANET in particular led to the development of protocols for internetworking, where multiple separate networks could be joined together into a network of networks. In 1982 the Internet Protocol Suite (TCP/IP) was standardized and the concept of a world-wide network of fully interconnected TCP/IP networks called the Internet was introduced. -
ARPANET and Its Boundary Devices: Modems, Imps, and the Inter-Structuralism of Infrastructures
Internet Histories Digital Technology, Culture and Society ISSN: 2470-1475 (Print) 2470-1483 (Online) Journal homepage: http://www.tandfonline.com/loi/rint20 ARPANET and its boundary devices: modems, IMPs, and the inter-structuralism of infrastructures Fenwick McKelvey & Kevin Driscoll To cite this article: Fenwick McKelvey & Kevin Driscoll (2018): ARPANET and its boundary devices: modems, IMPs, and the inter-structuralism of infrastructures, Internet Histories, DOI: 10.1080/24701475.2018.1548138 To link to this article: https://doi.org/10.1080/24701475.2018.1548138 Published online: 28 Dec 2018. Submit your article to this journal Article views: 2 View Crossmark data Full Terms & Conditions of access and use can be found at http://www.tandfonline.com/action/journalInformation?journalCode=rint20 INTERNET HISTORIES https://doi.org/10.1080/24701475.2018.1548138 ARPANET and its boundary devices: modems, IMPs, and the inter-structuralism of infrastructures Fenwick McKelveya and Kevin Driscollb aDepartment of Communication Studies, Concordia University, Montreal, Quebec, Canada; bMedia Studies, University of Virginia, Charlottesville, Virginia, USA ABSTRACT ARTICLE HISTORY Our paper focuses on the Interface Message Processor (IMP), an Received 21 April 2018 important device in the history of ARPANET. Designed as the Revised 1 November 2018 interface between ARPANET nodes and the common carrier tele- Accepted 2 November 2018 phone system, the IMP actualized the ARPANET as an experimen- KEYWORDS tal packet-switching communication system. We conceptualize ’ ARPANET; infrastructure; the IMP as historical boundary object that exposes ARPANET s boundary objects; gateways; close relationship to the telephone system. Our analysis offers a internetworking novel history of ARPANET as a repurposing of the existing tele- phone infrastructure. -
Notes on the History of Fork-And-Join Linus Nyman and Mikael Laakso Hanken School of Economics
Notes on the history of fork-and-join Linus Nyman and Mikael Laakso Hanken School of Economics As part of a PhD on code forking in open source software, Linus Nyman looked into the origins of how the practice came to be called forking.1 This search led back to the early history of the fork system call. Having not previously seen such a history published, this anecdote looks back at the birth of the fork system call to share what was learned, as remembered by its pioneers. The fork call allows a process (or running program) to create new processes. The original is deemed the parent process and the newly created one its child. On multiprocessor systems, these processes can run concurrently in parallel.2 Since its birth 50 years ago, the fork has remained a central element of modern computing, both with regards to software development principles and, by extension, to hardware design, which increasingly accommodates parallelism in process execution. The fork system call is imagined The year was 1962. Melvin Conway, later to become known for “Conway’s Law,”3 was troubled by what seemed an unnecessary inefficiency in computing. As Conway recalls:4 I was in the US Air Force at the time – involved with computer procurement – and I noticed that no vendor was distinguishing between “processor” and “process.” That is, when a proposal involved a multiprocessor design, each processor was committed to a single process. By late 1962, Conway had begun contemplating the idea of using a record to carry the status of a process. -
Admission Control for Independently-Authored Realtime Applications
Admission Control for Independently-authored Realtime Applications by Robert Kroeger Athesis presented to the University of Waterloo in fulfilment of the thesis requirement for the degree of Doctor of Philosophy in Computer Science Waterloo, Ontario, Canada, 2004 c Robert Kroeger 2004 AUTHOR’S DECLARATION FOR ELECTRONIC SUBMISSION OF A THESIS I hereby declare that I am the sole author of this thesis. This is a true copy of the thesis, including any required final revisions, as accepted by my examiners. I understand that my thesis may be made electronically available to the public. ii Abstract This thesis presents the LiquiMedia operating system architecture. LiquiMedia is specialized to schedule multimedia applications. Because they generate output for a human observer, multimedia applications such as video games, video conferencing and video players have both unique scheduling requirements and unique allowances: a multimedia stream must synchro- nize sub-streams generated for different sensory modalities within 20 milliseconds, it is not successfully segregated until it has existed for over 200 milliseconds and tolerates occasional scheduling failures. LiquiMedia is specialized around these requirements and allowances. First, LiquiMedia synchronizes multimedia tasks by invoking them from a shared realtime timer interrupt. Sec- ond, owing to multimedia’s tolerance of scheduling failures, LiquiMedia schedules tasks based on a probabilistic model of their running times. Third, LiquiMedia can infer per-task models while a user is segregating the streams that the tasks generate. These specializations provide novel capabilities: up to 2.5 times higher utilization than RMS scheduling, use of an atomic task primitive 9.5 times more efficient than preemptive threading, and most importantly, the ability to schedule arbitrary tasks to a known probability of realtime execution without a priori knowledge of their running times. -
Types for Safe Locking: Static Race Detection for Java
Types for Safe Locking: Static Race Detection for Java MARTIN ABADI Microsoft Research and University of California at Santa Cruz CORMAC FLANAGAN University of California at Santa Cruz and STEPHEN N. FREUND Williams College This article presents a static race-detection analysis for multithreaded shared-memory programs, focusing on the Java programming language. The analysis is based on a type system that captures many common synchronization patterns. It supports classes with internal synchronization, classes that require client-side synchronization, and thread-local classes. In order to demonstrate the effectiveness of the type system, we have implemented it in a checker and applied it to over 40,000 lines of hand-annotated Java code. We found a number of race conditions in the standard Java libraries and other test programs. The checker required fewer than 20 additional type annotations per 1,000 lines of code. This article also describes two improvements that facilitate checking much larger programs: an algorithm for annotation inference and a user interface that clarifies warnings generated by the checker. These extensions have enabled us to use the checker for identifying race conditions in large-scale software systems with up to 500,000 lines of code. Categories and Subject Descriptors: F.3.1 [Logics and Meanings of Programs]: Specifying and Verifying and Reasoning about Programs; D.2.4 [Software Engineering]: Software/Program Ver- ification—Reliability; D.3.2 [Programming Languages]: Language Classifications—Concurrent, distributed, and parallel languages; D.3.1 [Programming Languages]: Formal Definitions and Theory—Semantics, syntax Martin Abadi’s work was started at Compaq; it was also partly supported by the National Science Foundation under Grants CCR-0204162 and CCR-0341179, and by faculty research funds granted by the University of California, Santa Cruz. -
Luca Cardelli • Curriculum Vitae • 2014
Luca Cardelli • Curriculum Vitae • 2014 Short Bio Luca Cardelli was born near Montecatini Terme, Italy, studied at the University of Pisa (until 1978-07- 12), and has a Ph.D. in computer science from the University of Edinburgh(1982-04-01). He worked at Bell Labs, Murray Hill, from 1982-04-05 to 1985-09-20, and at Digital Equipment Corporation, Systems Research Center in Palo Alto, from 1985-09-30 to 1997-10-31, before assuming a position on 1997-11-03 at Microsoft Research, in Cambridge UK, where he was head of the Programming Principles and Tools andSecurity groups until 2012, and is currently a Principal Researcher. His main interests are in type theory and operational semantics (for applications to language design, semantics, and implementation), and in concurrency theory (for applications to computer networks and to modeling biological systems). He implemented the first compiler for ML (one of the most popular typed functional language, whose recent incarnations are Caml and F#) and one of the earliest direct-manipulation user-interface editors. He was a member of the Modula-3 design committee, and has designed a few experimental languages, including Obliq: a distributed higher- order scripting language (voted most influential POPL'95 paper 10 years later), and Polyphonic C#, a distributed extension of C#. His more protracted research activity has been in establishing the semantic and type-theoretic foundations of object-oriented languages, resulting in the 1996 book "A Theory of Objects" with Martin Abadi. More recently he has focused on modeling global and mobile computation, via the Ambient Calculus andSpatial Logics, which indirectly led to a current interest in Systems Biology , Molecular Programming, and Stochastic Systems.