Id Question Microprocessor Is the Example of ___Architecture. A

Total Page:16

File Type:pdf, Size:1020Kb

Id Question Microprocessor Is the Example of ___Architecture. A Id Question Microprocessor is the example of _______ architecture. A Princeton B Von Neumann C Rockwell D Harvard Answer A Marks 1 Unit 1 Id Question _______ bus is unidirectional. A Data B Address C Control D None of these Answer B Marks 1 Unit 1 Id Question Use of _______isolates CPU form frequent accesses to main memory. A Local I/O controller B Expansion bus interface C Cache structure D System bus Answer C Marks 1 Unit 1 Id Question _______ buffers data transfer between system bus and I/O controllers on expansion bus. A Local I/O controller B Expansion bus interface C Cache structure D None of these Answer B Marks 1 Unit 1 Id Question ______ Timing involves a clock line. A Synchronous B Asynchronous C Asymmetric D None of these Answer A Marks 1 Unit 1 Id Question -----timing takes advantage of mixture of slow and fast devices, sharing the same bus A Synchronous B asynchronous C Asymmetric D None of these Answer B Marks 1 Unit 1 Id st Question In 1 generation _____language was used to prepare programs. A Machine B Assembly C High level programming D Pseudo code Answer B Marks 1 Unit 1 Id Question The transistor was invented at ________ laboratories. A AT &T B AT &Y C AM &T D AT &M Answer A Marks 1 Unit 1 Id Question ______is used to control various modules in the system A Control Bus B Data Bus C Address Bus D All of these Answer A Marks 1 Unit 1 Id Question The disadvantage of single bus structure over multi bus structure is _____ A Propagation delay B Bottle neck because of bus capacity C Both A and B D Neither A nor B Answer C Marks 1 Unit 1 Id Question _____ provide path for moving data among system modules A Control Bus B Data Bus C Address Bus D All of these Answer B Marks 1 Unit 1 Id Question Microcontroller is the example of _____ architecture. A Princeton B Von Neumann C Rockwell D Harvard Answer D Marks 1 Unit 1 Id Question A computer with more than one CPU allows instruction from different program to be executed simultaneously is called as A Microprocessor B Miniprocessor C super processor D multiprocessor Answer D Marks 1 Unit 1 Id Question The compact combination of CPU , memory and IO circuits in one System is called A Microcomputer B Minicomputer C Supercomputer D None of above Answer A Marks 1 Unit 1 Id Question A computer system with one CPU or microprocessor is generally referred to____ A Mainframe B Microcomputer C Large computer D None of above Answer B Marks 1 Unit 1 Id Question Performance of system and execution time of system are ___ proportional to each other. A Directly B Inversely C Equally D None of the above Answer B Marks 1 Unit 1 Id Question In pipelined processing common steps involved in instruction processing are_____overlapped A partially B completely C Randomly D None of the above Answer A Marks 1 Unit 1 Id Question In non pipelined CPU, instructions are executed in___sequence A overlapped B any C fixed D None of the above Answer C Marks 1 Unit 1 Id Question Computers main or primary memory M is____ A ROM B RAM C PROM D None of the above Answer B Marks 1 Unit 1 Id Question Supercomputers mainly used in scientific calculations involving large amounts of vectors and matrix calculations are some times referred as___ A super processor B Vector processor C Parallel processor D Scientific processor Answer B Marks 1 Unit 1 Id Question A control signal can be ____ A Memory read B Memory write C I/ O read D All of these Answer D Marks 1 Id Question1 In Von Neumann Machine ___ feature was responsible for performance bottleneck. A Stored program B Separate memory for data and code C I/O access D None of these Answer A Marks 1 Unit 1 Id Question2 CPU Means ___ A Control Processing Unit B Control Programming Unit C Central Processing Unit D Central Programming Unit Answer C Marks 1 Unit 1 Id Question3 Key concept of stored program was introduced by____ A Safwat G. Zaky B John Von Neumann C j. Hays D Stalling William Answer B Marks 1 Unit 1 Id Question4 Which of the following operation are involved in an instruction cycle? A Opcode decoding B Instruction execution C Instruction fetching D All of these Answer D Marks 1 Unit 1 Id Question5 ___stores address of next instruction to be executed. A AR B AC C PC D IR Answer C Marks 1 Unit 1 Id Question6 ____keep track of execution of a program. A Program Counter B Instruction Register C General Purpose Register D Memory Address Register Answer A Marks 1 Unit 1 Id Question7 ___contains the data to be written into or read out of address location. A MAR B PC C MDR D IR Answer C Marks 1 Unit 1 Id Question8 In IAS actual word transfer take place between memory and ___ A AC B AR C DR D IR Answer C Marks 1 Unit 1 Id Question9 Data register of IAS is __wide. A 16-bit B 20-bit C 40-bit D 48-bit Answer C Marks 1 Unit 1 Id Question1 AR in IAS is ___ wide. 0 A 12-bit B 14-bit C 16-bit D 18-bit Answer A Marks 1 Unit 1 Id Question1 Program control unit of IAS fetches ____ instruction 1 simultaneously. A Two B Three C Four D Six Answer A Marks 1 Unit 1 Id Question1 In IAS machine, IBR stores _____ instruction. 2 A Immediately executable B Later executable C Currently executing D Aborted type Answer B Marks 1 Unit 1 Id Questions In IAS machine, IR stores _____ instruction. 13 A Immediately executable B Later executable C Currently executing D Aborted type Answer A Marks 1 Unit 1 Id Question1 _______ Architecture shows separate memory banks for 4 data and programs. A Princeton B Harvard C Von Neumann D Rockwell Answer B Marks 1 Unit 1 Id Question1 Harvard architecture shows features of executing 5 instruction in _______ instruction cycle than von Neumann. A More B Double C Reduced D Exactly half Answer C Marks 1 Unit 1 Id Question1 ________ Architecture uses both RISC and CISC 6 architectures. A Von Neumann B Harvard C Babbage D None of these Answer A Marks 1 Unit 1 Id Question1 In IAS machine, program control unit fetches………….. 7 instructions simultaneously from memory A One B Two C Three D four Answer B Marks 1 Unit 1 Id Question1 IAS machines are designed to process all bits of binary 8 numbers….. A Randomly B Simultaneously C Serially D None of the above Answer B Marks 1 Unit 1 Id Question1 In CPU of IAS computer who issues control signals to 9 data processing unit, memory and other circuits for execution of instruction? A Program Counter PC B Program Control Unit PCU C Accumulator AC D Instruction Register IR Answer B Marks 1 Unit 1 Id Question2 In IAS Computer ____fetches and interprets the 0 instruction in memory and causes them to be executed. A ALU B Control Unit C Accumulator D Program counter Answer B Marks 1 Unit 1 Id Question2 In original Von Neumann Machine, memory unit consists 1 of ___ storage location of __bits each. A 2048,16 B 4096,20 C 4096,60 D 4096,40 Answer D Marks 2 Unit 1 Id Question2 In Original IAS Machine ,storage location was referred to 2 as ____ A Byte B Word C Digit D Number Answer B Marks 1 Unit 1 Id Question2 In IAS machine, data and code have _____memory. 3 A Same B Separate C Distinct D More Answer A Marks 1 Unit 1 Id Question1 The three main parts of a stored programmed computer are A CPU , memory, I/ O B Register, memory control unit C CPU, register, ALU D CPU, I/ O , control unit Answer A Marks 1 Unit 1 Id Question2 CPU takes___ time to obtain word from memory than one of its internal register A Same B longer C Lesser D Zero Answer B Marks 1 Unit 1 Id Question3 If performance of system A is 2times the performance of system B then execution times for system A is ____as of execution time for system B. A half B 2 times C same D None of above Answer A Marks 1 Unit 1 Id Question4 To enhance the speed of processor, small memory unit placed between CPU and main memory M is called_______ A Program memory B Stack memory C Cache memory D Dynamic memory Answer C Marks 1 Unit 1 Id Question5 Cache memory can store_______ A Only instructions B Instructions and data C Only data D None of the above Answer B Marks 1 Unit 1 Id Question6 Cache memory is usually___than main memory. A bigger B smaller C faster D (B) and (C) both Answer D Marks 1 Unit 1 Id Question7 Most I/O operation are ___memory operation A easier B faster C slower D None of the above Answer C Marks 1 Unit 1 Id Question8 Major component of most I/O system is set of ___memory devices A secondary B primary C cache D None of the above Answer A Marks 1 Unit 1 Id Question9 I/O devices are attached to host computer by means of___ A I/O ports B Memory slots C Control bus D None of the above Answer A Marks 1 Unit 1 Id Question1 In early supercomputer parallel processing was heavily 0 depend on___ A Sequential processing B Random processing C Pipeline processing D Nonpipeline processing Answer C Marks 1 Unit 1 Id Question Collection of data, address and control Bus is usually 1 referred as____.
Recommended publications
  • Computer Organization and Architecture Designing for Performance Ninth Edition
    COMPUTER ORGANIZATION AND ARCHITECTURE DESIGNING FOR PERFORMANCE NINTH EDITION William Stallings Boston Columbus Indianapolis New York San Francisco Upper Saddle River Amsterdam Cape Town Dubai London Madrid Milan Munich Paris Montréal Toronto Delhi Mexico City São Paulo Sydney Hong Kong Seoul Singapore Taipei Tokyo Editorial Director: Marcia Horton Designer: Bruce Kenselaar Executive Editor: Tracy Dunkelberger Manager, Visual Research: Karen Sanatar Associate Editor: Carole Snyder Manager, Rights and Permissions: Mike Joyce Director of Marketing: Patrice Jones Text Permission Coordinator: Jen Roach Marketing Manager: Yez Alayan Cover Art: Charles Bowman/Robert Harding Marketing Coordinator: Kathryn Ferranti Lead Media Project Manager: Daniel Sandin Marketing Assistant: Emma Snider Full-Service Project Management: Shiny Rajesh/ Director of Production: Vince O’Brien Integra Software Services Pvt. Ltd. Managing Editor: Jeff Holcomb Composition: Integra Software Services Pvt. Ltd. Production Project Manager: Kayla Smith-Tarbox Printer/Binder: Edward Brothers Production Editor: Pat Brown Cover Printer: Lehigh-Phoenix Color/Hagerstown Manufacturing Buyer: Pat Brown Text Font: Times Ten-Roman Creative Director: Jayne Conte Credits: Figure 2.14: reprinted with permission from The Computer Language Company, Inc. Figure 17.10: Buyya, Rajkumar, High-Performance Cluster Computing: Architectures and Systems, Vol I, 1st edition, ©1999. Reprinted and Electronically reproduced by permission of Pearson Education, Inc. Upper Saddle River, New Jersey, Figure 17.11: Reprinted with permission from Ethernet Alliance. Credits and acknowledgments borrowed from other sources and reproduced, with permission, in this textbook appear on the appropriate page within text. Copyright © 2013, 2010, 2006 by Pearson Education, Inc., publishing as Prentice Hall. All rights reserved. Manufactured in the United States of America.
    [Show full text]
  • Parallel Computing
    Lecture 1: Computer Organization 1 Outline • Overview of parallel computing • Overview of computer organization – Intel 8086 architecture • Implicit parallelism • von Neumann bottleneck • Cache memory – Writing cache-friendly code 2 Why parallel computing • Solving an × linear system Ax=b by using Gaussian elimination takes ≈ flops. 1 • On Core i7 975 @ 4.0 GHz,3 which is capable of about 3 60-70 Gigaflops flops time 1000 3.3×108 0.006 seconds 1000000 3.3×1017 57.9 days 3 What is parallel computing? • Serial computing • Parallel computing https://computing.llnl.gov/tutorials/parallel_comp 4 Milestones in Computer Architecture • Analytic engine (mechanical device), 1833 – Forerunner of modern digital computer, Charles Babbage (1792-1871) at University of Cambridge • Electronic Numerical Integrator and Computer (ENIAC), 1946 – Presper Eckert and John Mauchly at the University of Pennsylvania – The first, completely electronic, operational, general-purpose analytical calculator. 30 tons, 72 square meters, 200KW. – Read in 120 cards per minute, Addition took 200µs, Division took 6 ms. • IAS machine, 1952 – John von Neumann at Princeton’s Institute of Advanced Studies (IAS) – Program could be represented in digit form in the computer memory, along with data. Arithmetic could be implemented using binary numbers – Most current machines use this design • Transistors was invented at Bell Labs in 1948 by J. Bardeen, W. Brattain and W. Shockley. • PDP-1, 1960, DEC – First minicomputer (transistorized computer) • PDP-8, 1965, DEC – A single bus
    [Show full text]
  • P the Pioneers and Their Computers
    The Videotape Sources: The Pioneers and their Computers • Lectures at The Compp,uter Museum, Marlboro, MA, September 1979-1983 • Goal: Capture data at the source • The first 4: Atanasoff (ABC), Zuse, Hopper (IBM/Harvard), Grosch (IBM), Stibitz (BTL) • Flowers (Colossus) • ENIAC: Eckert, Mauchley, Burks • Wilkes (EDSAC … LEO), Edwards (Manchester), Wilkinson (NPL ACE), Huskey (SWAC), Rajchman (IAS), Forrester (MIT) What did it feel like then? • What were th e comput ers? • Why did their inventors build them? • What materials (technology) did they build from? • What were their speed and memory size specs? • How did they work? • How were they used or programmed? • What were they used for? • What did each contribute to future computing? • What were the by-products? and alumni/ae? The “classic” five boxes of a stored ppgrogram dig ital comp uter Memory M Central Input Output Control I O CC Central Arithmetic CA How was programming done before programming languages and O/Ss? • ENIAC was programmed by routing control pulse cables f ormi ng th e “ program count er” • Clippinger and von Neumann made “function codes” for the tables of ENIAC • Kilburn at Manchester ran the first 17 word program • Wilkes, Wheeler, and Gill wrote the first book on programmiidbBbbIiSiing, reprinted by Babbage Institute Series • Parallel versus Serial • Pre-programming languages and operating systems • Big idea: compatibility for program investment – EDSAC was transferred to Leo – The IAS Computers built at Universities Time Line of First Computers Year 1935 1940 1945 1950 1955 ••••• BTL ---------o o o o Zuse ----------------o Atanasoff ------------------o IBM ASCC,SSEC ------------o-----------o >CPC ENIAC ?--------------o EDVAC s------------------o UNIVAC I IAS --?s------------o Colossus -------?---?----o Manchester ?--------o ?>Ferranti EDSAC ?-----------o ?>Leo ACE ?--------------o ?>DEUCE Whirl wi nd SEAC & SWAC ENIAC Project Time Line & Descendants IBM 701, Philco S2000, ERA..
    [Show full text]
  • Parallel Computing
    Lecture 1: Single processor performance Why parallel computing • Solving an 푛 × 푛 linear system Ax=b by using Gaussian 1 elimination takes ≈ 푛3 flops. 3 • On Core i7 975 @ 4.0 GHz, which is capable of about 60-70 Gigaflops 푛 flops time 1000 3.3×108 0.006 seconds 1000000 3.3×1017 57.9 days Milestones in Computer Architecture • Analytic engine (mechanical device), 1833 – Forerunner of modern digital computer, Charles Babbage (1792-1871) at University of Cambridge • Electronic Numerical Integrator and Computer (ENIAC), 1946 – Presper Eckert and John Mauchly at the University of Pennsylvania – The first, completely electronic, operational, general-purpose analytical calculator. 30 tons, 72 square meters, 200KW. – Read in 120 cards per minute, Addition took 200µs, Division took 6 ms. • IAS machine, 1952 – John von Neumann at Princeton’s Institute of Advanced Studies (IAS) – Program could be represented in digit form in the computer memory, along with data. Arithmetic could be implemented using binary numbers – Most current machines use this design • Transistors was invented at Bell Labs in 1948 by J. Bardeen, W. Brattain and W. Shockley. • PDP-1, 1960, DEC – First minicomputer (transistorized computer) • PDP-8, 1965, DEC – A single bus (omnibus) connecting CPU, Memory, Terminal, Paper tape I/O and Other I/O. • 7094, 1962, IBM – Scientific computing machine in early 1960s. • 8080, 1974, Intel – First general-purpose 8-bit computer on a chip • IBM PC, 1981 – Started modern personal computer era Remark: see also http://www.computerhistory.org/timeline/?year=1946
    [Show full text]
  • Building Computers in 1953: JOHNNIAC"
    The Computer Museum History Center presents "Building Computers in 1953: JOHNNIAC" Willis Ware Bill Gunning JOHNNIAC Designer JOHNNIAC Project Engineer Paul Armer Mort Bernstein RAND Department Head JOHNNIAC Software Developer 5:30 PM, Tuesday, Sept. 15, 1998 Computer Museum History Center Building 126 Moffett Field Mt. View JOHNNIAC was one of an illustrious group of computers built in the early 1950's, all inspired by the IAS computer designed by John von Neumann at the Institute for Advanced Study in Princeton. Some of these machines were MANIAC (Los Alamos) and ILLIAC (Univ. of Illinois), as well as WEIZAC, AVIDAC, and ORDVAC. JOHNNIAC was built at RAND Corporation in Santa Monica, and named after John von Neumann himself (he "disapproved", but didn't complain too much). This talk will be given in front of JOHNNIAC itself, the actual machine, since it is now part of the Computer Museum History Center at Moffett Field, the world's largest collection of historical computer hardware. JOHNNIAC ran for the first time in March, 1954. It pioneered the development of time shared operating systems with JOSS (the JOHNNIAC Open Shop System). JOSS could support several dozen users with drum swapping -- there were no disk drives then, so no disk swapping, but instead large rotating drums for magnetic storage. JOHNNIAC also inaugurated the commercial use of magnetic core memory, iron doughnuts on wires, which dominated computer memories for the next decade. Among other tasks, JOHNNIAC was also used to develop digitizing tablets for computer input. Our speakers were all working on JOHNNIAC over 40 years ago.
    [Show full text]
  • The Baleful Effect of Computer Languages and Benchmarks Upon Applied Mathematics, Physics and Chemistry
    SIAMjvnl July 16, 1997 The John von Neumann Lecture on The Baleful Effect of Computer Languages and Benchmarks upon Applied Mathematics, Physics and Chemistry presented by Prof. W. Kahan Mathematics Dept., and Elect. Eng. & Computer Sci. Dept. University of California Berkeley CA 94720-1776 (510) 642-5638 at the 45th Annual Meeting of S.I.A.M. Stanford University 15 July 1997 PostScript file: http://http.cs.berkeley.edu/~wkahan/SIAMjvnl.ps For more details see also …/Triangle.ps and …/Cantilever.ps Slide 1 SIAMjvnl July 16, 1997 The Baleful Effect of Computer Languages and Benchmarks upon Applied Mathematics, Physics and Chemistry Abstract: An unhealthy preoccupation with floating–point arithmetic’s Speed, as if it were the same as Throughput, has distracted the computing industry and its marketplace from other important qualities that computers’ arithmetic hardware and software should possess too, qualities like Accuracy, Reliability, Ease of Use, Adaptability, ··· These qualities, especially the ones that arise out of Intellectual Economy, tend to be given little weight partly because they are so hard to quantify and to compare meaningfully. Worse, technical and political limitations built into current benchmarking practices discourage innovations and accommodations of features necessary as well as desirable for robust and reliable technical computation. Most exasperating are computer languages like Java that lack locutions to access advantageous features in hardware that we consequently cannot use though we have paid for them. That lack prevents benchmarks from demonstrating the features' benefits, thus denying language implementors any incentive to accommodate those features in their compilers. It is a vicious circle that the communities concerned with scientific and engineering computation must help the computing industry break.
    [Show full text]
  • Computer Architecture ELE 475 / COS 475 Slide Deck 1
    Computer Architecture ELE 475 / COS 475 Slide Deck 1: Introduction and Instruction Set Architectures David Wentzlaff Department of Electrical Engineering Princeton University 1 What is Computer Architecture? Application 2 What is Computer Architecture? Application Physics 3 What is Computer Architecture? Application Gap too large to bridge in one step Physics 4 What is Computer Architecture? Application In its broadest definition, Gap too large to computer architecture is the design of the abstraction/ bridge in one step implementation layers that allow us to execute information processing applications efficiently using manufacturing technologies Physics 5 What is Computer Architecture? Application In its broadest definition, Gap too large to computer architecture is the design of the abstraction/ bridge in one step implementation layers that allow us to execute information processing applications efficiently using manufacturing technologies Physics 6 Abstractions in Modern Computing Systems Application Algorithm Programming Language Operating System/Virtual Machines Instruction Set Architecture Microarchitecture Register-Transfer Level Gates Circuits Devices Physics 7 Abstractions in Modern Computing Systems Application Algorithm Programming Language Operating System/Virtual Machines Instruction Set Architecture Computer Architecture Microarchitecture (ELE 475) Register-Transfer Level Gates Circuits Devices Physics 8 Computer Architecture is Constantly Changing Application Application Requirements: Algorithm • Suggest how to improve
    [Show full text]
  • 1.12 Historical Perspective and Further Reading 54.E1
    1.12 Historical Perspective and Further Reading 54.e1 An active fi eld of science is like an 1.12 Historical Perspective and Further immense anthill; the Reading individual almost vanishes into the mass of minds tumbling over For each chapter in the text, a section devoted to a historical perspective can each other, carrying be found online. We may trace the development of an idea through a series of information from place machines or describe some important projects, and we provide references in case to place, passing it you are interested in probing further. around at the speed of Th e historical perspective for this chapter provides a background for some of the light. key ideas presented therein. Its purpose is to give you the human story behind the Lewis Th omas, “Natural technological advances and to place achievements in their historical context. By Science,” in Th e Lives of learning the past, you may be better able to understand the forces that will shape a Cell , 1974 computing in the future. Each historical perspective section ends with suggestions for additional reading, which are also collected separately in the online section “Further Reading.” The First Electronic Computers J. Presper Eckert and John Mauchly at the Moore School of the University of Pennsylvania built what is widely accepted to be the world’s fi rst operational electronic, general-purpose computer. Th is machine, called ENIAC ( Electronic Numerical Integrator and Calculator), was funded by the United States Army and started working during World War II but was not publicly disclosed until 1946.
    [Show full text]
  • Assembly Language
    x86 Architectures; Assembly Language Basics of Assembly language for the x86 and x86_64 architectures topics • Preliminary material – a look at what Assembly Language works with - How processors work » A moment of history - Programming model » the registers » the memory The Processor and The System Unit Computer Organization – Programmer's View • Fundamental components: - CPU - Memory - I/O (“everything else”) • Possible connection schemes: - Single Controller (right) - Direct Memory Access (DMA) (below) the Von Neumann Central Processing Unit • a.k.a. CPU or Processor Control - (or microprocessor) Unit • Control unit interprets program instructions Arithmetic • ALU performs numerical Logic and logical operations Unit • Registers access program data Registers • von Neumann model: single connection to memory to/from - the "Von Neumann Bottleneck" memory John von Neumann’s IAS Machine • Built at the Institute for Advanced Studies, Princeton, NJ - 1945 - 1951 • Program instructions and data in single shared memory - ~5 Kilobytes Data Flow 1. Instructions move from memory (through cache) into Control Unit 2. Operands move between memory and data registers 3. Instruction execution causes data movement from/to memory (sometimes) - Computations carry operands from registers (or memory) to ALU, FPU, then back to registers (or memory) - Input/Output operations move values between devices and registers, or between devices and memory locations Modern CPU elements GPU • ALU, FPU, GPU - Do the computations • General-Purpose Data Arithmetic- Registers Registers Logic Units - Hold data for ALU, FPU Floating- • Specialized Registers Point Units - Control Program Flow Specialized Registers • Control Unit - Controls everything Control Unit Memory • Internal busses (a.k.a. - Connect things Store, Storage, • Cache cache or RAM) - Alleviates Von Neumann bottleneck - Generally not visible to programmer Multi-core Processors • Each "core" is a CPU • One chip holds 2, 4, 8, ..
    [Show full text]
  • Appendix M Historical Perspectives and References
    M.1 Introduction M-2 M.2 The Early Development of Computers (Chapter 1) M-2 M.3 The Development of Memory Hierarchy and Protection (Chapter 2 and Appendix B) M-9 M.4 The Evolution of Instruction Sets (Appendices A, J, and K) M-17 M.5 The Development of Pipelining and Instruction-Level Parallelism (Chapter 3 and Appendices C and H) M-27 M.6 The Development of SIMD Supercomputers, Vector Computers, Multimedia SIMD Instruction Extensions, and Graphical Processor Units (Chapter 4) M-45 M.7 The History of Multiprocessors and Parallel Processing (Chapter 5 and Appendices F, G, and I) M-55 M.8 The Development of Clusters (Chapter 6) M-74 M.9 Historical Perspectives and References M-79 M.10 The History of Magnetic Storage, RAID, and I/O Buses (Appendix D) M-84 M Historical Perspectives and References If … history … teaches us anything, it is that man in his quest for knowledge and progress is determined and cannot be deterred. John F. Kennedy Address at Rice University (1962) Those who cannot remember the past are condemned to repeat it. George Santayana The Life of Reason (1905), Vol. 2, Chapter 3 M-2 ■ Appendix M Historical Perspectives and References M.1 Introduction This appendix provides historical background on some of the key ideas presented in the chapters. We may trace the development of an idea through a series of machines or describe significant projects. If you are interested in examining the initial development of an idea or machine or are interested in further reading, references are provided at the end of each section.
    [Show full text]
  • William Stallings Computer Organization and Architecture 8Th Edition
    William Stallings Computer Organization and Architecture 8th Edition Chapter 1 Introduction Architecture & Organization 1 • Architecture is those attributes visible to the programmer —Instruction set, number of bits used for data representation, I/O mechanisms, addressing techniques. —e.g. Is there a multiply instruction? • Organization is how features are implemented —Control signals, interfaces, memory technology. —e.g. Is there a hardware multiply unit or is it done by repeated addition? Architecture & Organization 2 • All Intel x86 family share the same basic architecture • The IBM System/370 family share the same basic architecture • This gives code compatibility ... at least backwards • Organization differs within members of the same family, e.g. floating point numerical co- processors with names like 8087, 80287 and 80387. With very few exceptions, the 80486 and subsequent x86 processors then integrated this x87 functionality on chip. Structure & Function • Structure is the way in which components relate to each other • Function is the operation of individual components as part of the structure Function General computer functions: —Data processing —Data storage —Data movement —Control Operations (a) Data movement I/O (peripherals directly attached) Communications/Networking (communication lines) Example application? Operations (a) Data movement I/O (peripherals directly attached) Communications/Networking (communication lines) Camera attached to a PC, sending the frames to a window on the screen of the same PC. Operations (b) Storage Example application? Operations (b) Storage Playing an mp3 file stored in memory to earphones attached to the same PC. Operation (c) Processing from/to storage Example application? Operation (c) Processing from/to storage Any number-crunching application that takes data from memory and stores the result back in memory.
    [Show full text]
  • 1.12 Historical Perspective and Further Reading 1.12-3 High and Several Feet Wide
    An active fi eld of science is like an immense anthill; the 1.12 Historical Perspective and Further individual almost Reading vanishes into the mass of minds tumbling over each other, carrying For each chapter in the text, a section devoted to a historical perspective can information from place be found online. We may trace the development of an idea through a series of to place, passing it machines or describe some important projects, and we provide references in case around at the speed of you are interested in probing further. light. Th e historical perspective for this chapter provides a background for some of the Lewis Th omas, “Natural key ideas presented therein. Its purpose is to give you the human story behind the Science,” in Th e Lives of technological advances and to place achievements in their historical context. By a Cell , 1974 learning the past, you may be better able to understand the forces that will shape computing in the future. Each historical perspective section ends with suggestions for additional reading, which are also collected separately in the online section “Further Reading.” The First Electronic Computers J. Presper Eckert and John Mauchly at the Moore School of the University of Pennsylvania built what is widely accepted to be the world’s fi rst operational electronic, general-purpose computer. Th is machine, called ENIAC ( Electronic Numerical Integrator and Calculator), was funded by the United States Army and started working during World War II but was not publicly disclosed until 1946. ENIAC was a general-purpose machine used for computing artillery-fi ring tables.
    [Show full text]