Part IV: Turing Machines and Undecidability

Total Page:16

File Type:pdf, Size:1020Kb

Part IV: Turing Machines and Undecidability Part IV: Turing Machines and Undecidability We are about to begin our exploration of the two outer circles of the language hierarchy, as well as the background, the area outside all of the circles. Up until now, we have been placing limitations on what we could do in order that we could discover simple solutions when they exist. Now we are going to tear down all the barriers and explore the full power of formal computation. We will discover a whole range of problems that become solvable once we do that. We will also discover that there are fundamental limitations on what we can compute, regardless of the specific model with which we choose to work. Part IV 264 Turing Machines and Undecidability 17 Turing Machines We need a new kind of automaton that has two properties: • It must be powerful enough to describe all computable things. In this respect, it should be like real computers and unlike FSMs and PDAs. • It must be simple enough that we can reason formally about it. In this respect, it should be like FSMs and PDAs and unlike real computers. 17.1 Definition, Notation and Examples In our discussion of pushdown automata, it became clear that a finite state controller augmented by a single stack was not powerful enough to be able to execute even some very simple programs. What else must be added in order to acquire the necessary power? One answer is a second stack. We will explore that idea in Section 17.5.2. 17.1.1 What Is a Turing Machine? A more straightforward approach is to eliminate the stack and replace it by a more flexible form of infinite storage, a writeable tape. When we do that, we get a Turing machine . Figure 17. 1 showns a simple schematic diagram of a Turing machine M. M’s tape is infinite in both directions. The input to M is written on the tape, one character per square, before M is started. All other squares of the tape are initially blank (). As we have done for both FSMs and PDAs, M’s behavior will be defined only on input strings that are: • finite, and • contain only characters in M’s input alphabet. … a a b b b … Finite State Controller s, q1, q2, … h1, h2 Figure 17.1 The structure of a Turing machine M has a single read/write head, shown here with an arrow. We will almost always use the convention that, when M starts, its read/write head will be over the blank immediately to the left of the leftmost character of the input. But occasionally, when we are designing a machine to be used as a subroutine by some other machine, we may choose a different initial specification. M begins in its start state. At each step of its operation, M must: • choose its next state, • write on the current square, and • move the read/write head left or right one square. M can move back and forth on its tape, so there is no longer the idea that it consumes all of its input characters one at a time and then halts. M will continue to execute until it reaches a special state called a halting state. It is possible that M may never reach a halting state, in which case it will execute forever. Chapter 17 265 Turing Machines Notice that there will always be a finite number of nonblank squares on M’s tape. This follows from the fact that, before M starts, only a finite number of squares are nonblank. Then, at each step of its operation, M can write on at most one additional square. So, after any finite number of steps, only a finite number of squares can be nonblank. And, even if M never halts, at any point in its computation it will have executed only a finite number of steps. In Chapter 18 we are going to argue that the Turing machine, as we have just described it, is as powerful as any other reasonable model of computation, including modern computers. So, in the rest of this discussion, although our examples will be simple, remember that we are now talking about computation, broadly conceived. We are now ready to provide a formal definition: a Turing machine (or TM) M is a sixtuple (K, , , , s, H): • K is a finite set of states, • is the input alphabet, which does not contain , • is the tape alphabet, which must, at a minimum, contain and have as a subset, • s K is the start state, • H K is the set of halting states, and • is the transition function. It maps from: (K - H) to K {→, }. non-halting state tape character state tape character action (R or L) If contains the transition ((q0, a), (q1, b, A)) then, whenever M is in state q0 and the character under the read/write head is a, M will go to state q1, write b, and then move the read/write head as specified by A (either one square to the right or one square to the left). Notice that the tape symbol is special in two ways: • Initially, all tape squares except those that contain the input string contain . • The input string may not contain . But those are the only ways in which is special. A Turing machine may write just as it writes any other symbol in its tape alphabet . Be careful, though, if you design a Turing machine M that does write . Make sure that M can tell the difference between running off the end of the input and hitting a patch of ’s within the part of the tape it is working on. Some books use a definition of a Turing machine that does not allow writing . But we allow it because it can be quite useful if you are careful. In addition, this definition allows a Turing machine to output a string that is shorter than its input by writing ’s as necessary. Define the active tape of a Turing machine M to be the shortest fragment of M’s tape that includes the square under the read/write head and all the nonblank squares. We require that be defined for all (state, input) pairs unless the state is a halting state. Notice that is a function, not a relation. So this is a definition for deterministic Turing machines. One other important observation: a Turing machine can produce output, namely the contents of its tape when it halts. Example 17.1 Add b’s to Make Them Match the a’s Design a Turing machine M that takes as input a string in the language {aibj : 0 j i} and adds b's as required to make the number of b's equal the number of a's. The input to M will look like this: … a a a b … Chapter 17 266 Turing Machines On that input, the output (the contents of the tape when M halts) should be: … a a a b b b … M will operate as follows: 1. Move one square to the right. If the character under the read/write head is , halt. Otherwise, continue. 2. Loop: 2.1 Mark off an a with a $. 2.2 Scan rightward to the first b or . If b, mark it off with a # and get ready to go back and find the next matching a, b pair. If , then there are no more b’s but there are still a’s that need matches. So it is necessary to write another b on the tape. But that b must be marked so that it cannot match another a. So write a #. Then get ready to go back and look for remaining unmarked a’s. 2.3 Scan back leftward looking for a or . If a, then go back to the top of the loop and repeat. If , then all a’s have been handled. Exit the loop. (Notice that the specification for M guarantees that there will not be more b’s than a’s.) 3. Make one last pass all the way through the nonblank area of the tape, from left to right, changing each $ to an a and each # to a b. 4. Halt. M = {{1, 2, 3, 4, 5, 6}, {a, b}, {a, b, , $, #}, , 1, {6}), where = ( ((1, ), (2,, →)), ((1, a), (2,, →)), {These four transitions are required because M must be ((1, b), (2,, →)), defined for every state/ input pair, but since it isn’t possible ((1, $), (2,, →)), to see anything except in state 1, it doesn’t matter what ((1, #), (2,, →)), they do. } ((2, ), (6, $, →)), ((2, a), (3, $, →)), ((2, b), (3, $, →)), {Three more unusable elements of . We’ll omit the rest ((2, $), (3, $, →)), here for clarity.} ((2, #), (3, $, →)), ((3, ), (4, # , )), ((3, a), (3, a, →)), ((3, b), (4, #,)), ((3, $), (3, $, →)), ((3, #), (3, #, →)), ((4, ), (5, , →)), ((4, a), (3, $, →)), ((4, $), (4, $, )), ((4, #), (4, # , )), ((5, ), (6, ,)), {State 6 is a halting state and so has no transitions out of it.} ((5, $), (5, a, →)), ((5, #), (5, b , →)) ) People find it nearly impossible to read transition tables like this one, even for very simple machines. So we will adopt a graphical notation similar to the one we used for both FSMs and PDAs. Since each element of has five components, we need a notation for labeling arcs that includes all the required information. Let x/t/a on an arc of M mean that the transition can be taken if the character currently under the read/write head is x. If it is taken, write t and then move the read/write head as specified by a.
Recommended publications
  • PERL – a Register-Less Processor
    PERL { A Register-Less Processor A Thesis Submitted in Partial Fulfillment of the Requirements for the Degree of Doctor of Philosophy by P. Suresh to the Department of Computer Science & Engineering Indian Institute of Technology, Kanpur February, 2004 Certificate Certified that the work contained in the thesis entitled \PERL { A Register-Less Processor", by Mr.P. Suresh, has been carried out under my supervision and that this work has not been submitted elsewhere for a degree. (Dr. Rajat Moona) Professor, Department of Computer Science & Engineering, Indian Institute of Technology, Kanpur. February, 2004 ii Synopsis Computer architecture designs are influenced historically by three factors: market (users), software and hardware methods, and technology. Advances in fabrication technology are the most dominant factor among them. The performance of a proces- sor is defined by a judicious blend of processor architecture, efficient compiler tech- nology, and effective VLSI implementation. The choices for each of these strongly depend on the technology available for the others. Significant gains in the perfor- mance of processors are made due to the ever-improving fabrication technology that made it possible to incorporate architectural novelties such as pipelining, multiple instruction issue, on-chip caches, registers, branch prediction, etc. To supplement these architectural novelties, suitable compiler techniques extract performance by instruction scheduling, code and data placement and other optimizations. The performance of a computer system is directly related to the time it takes to execute programs, usually known as execution time. The expression for execution time (T), is expressed as a product of the number of instructions executed (N), the average number of machine cycles needed to execute one instruction (Cycles Per Instruction or CPI), and the clock cycle time (), as given in equation 1.
    [Show full text]
  • Computer Architecture Basics
    V03: Computer Architecture Basics Prof. Dr. Anton Gunzinger Supercomputing Systems AG Technoparkstrasse 1 8005 Zürich Telefon: 043 – 456 16 00 Email: [email protected] Web: www.scs.ch Supercomputing Systems AG Phone +41 43 456 16 00 Technopark 1 Fax +41 43 456 16 10 8005 Zürich www.scs.ch Computer Architecture Basic 1. Basic Computer Architecture Diagram 2. Basic Computer Architectures 3. Von Neumann versus Harvard Architecture 4. Computer Performance Measurement 5. Technology Trends 2 Zürich 20.09.2019 © by Supercomputing Systems AG CONFIDENTIAL 1.1 Basic Computer Architecture Diagram Sketch the basic computer architecture diagram Describe the functions of the building blocks 3 Zürich 20.09.2019 © by Supercomputing Systems AG CONFIDENTIAL 1.2 Basic Computer Architecture Diagram Describe the execution of a single instruction 5 Zürich 20.09.2019 © by Supercomputing Systems AG CONFIDENTIAL 2 Basic Computer Architecture 2.1 Sketch an Accumulator Machine 2.2 Sketch a Register Machine 2.3 Sketch a Stack Machine 2.4 Sketch the analysis of computing expression in a Stack Machine 2.5 Write the micro program for an Accumulator, a Register and a Stack Machine for the instruction: C:= A + B and estimate the numbers of cycles 2.6 Write the micro program for an Accumulator, a Register and a Stack Machine for the instruction: C:= (A - B)2 + C and estimate the numbers of cycles 2.7 Compare Accumulator, Register and Stack Machine 7 Zürich 20.09.2019 © by Supercomputing Systems AG CONFIDENTIAL 2.1 Accumulator Machine Draw the Accumulator Machine 8 Zürich
    [Show full text]
  • Computers, Complexity, and Controversy
    Instruction Sets and Beyond: Computers, Complexity, and Controversy Robert P. Colwell, Charles Y. Hitchcock m, E. Douglas Jensen, H. M. Brinkley Sprunt, and Charles P. Kollar Carnegie-Mellon University t alanc o tyreceived Instruction set design is important, but lt1Mthi it should not be driven solely by adher- ence to convictions about design style, ,,tl ica' ch m e n RISC or CISC. The focus ofdiscussion o09 Fy ipt issues. RISC should be on the more general question of the assignment of system function- have guided ality to implementation levels within years. A study of an architecture. This point of view en- d yield a deeper un- compasses the instruction set-CISCs f hardware/software tend to install functionality at lower mputer performance, the system levels than RISCs-but also JOCUS on rne assignment iluence of VLSI on processor design, takes into account other design fea- ofsystem functionality to and many other topics. Articles on tures such as register sets, coproces- RISC research, however, often fail to sors, and caches. implementation levels explore these topics properly and can While the implications of RISC re- within an architecture, be misleading. Further, the few papers search extend beyond the instruction and not be guided by that present comparisons with com- set, even within the instruction set do- whether it is a RISC plex instruction set computer design main, there are limitations that have or CISC design. often do not address the same issues. not been identified. Typical RISC As a result, even careful study of the papers give few clues about where the literature is likely to give a distorted RISC approach might break down.
    [Show full text]
  • A Systematic Study of Groebner Basis Methods
    A Systematic Study of Gr¨obner Basis Methods Vom Fachbereich Informatik der Technischen Universit¨at Kaiserslautern genehmigte Habilitationsschrift von Dr. Birgit Reinert Datum der Einreichung: 6. Januar 2003 Datum des wissenschaftlichen Vortrags: 9. Februar 2004 arXiv:0903.2462v2 [math.RA] 29 Mar 2009 Dekan: Prof. Dr. Hans Hagen Habilitationskommission: Vorsitzender: Prof. Dr. Otto Mayer Berichterstatter: Prof. Dr. Klaus E. Madlener Prof. Dr. Teo Mora Prof. Dr. Volker Weispfenning Vorwort Die vorliegende Arbeit ist die Quintessenz meiner Ideen und Erfahrungen, die ich in den letzten Jahren bei meiner Forschung auf dem Gebiet der Gr¨obnerbasen gemacht habe. Meine geistige Heimat war dabei die Arbeitsgruppe von Profes- sor Klaus Madlener an der Technischen Universit¨at Kaiserslautern. Hier habe ich bereits im Studium Bekanntschaft mit der Theorie der Gr¨obnerbasen gemacht und mich w¨ahrend meiner Promotion mit dem Spezialfall dieser Theorie f¨ur Monoid- und Gruppenringe besch¨aftigt. Nach der Promotion konnte ich im Rahmen eines DFG-Forschungsstipendiums zus¨atzlich Problemstellungen und Denkweisen an- derer Arbeitsgruppen kennenlernen - die Arbeitsgruppe von Professor Joachim Neub¨user in Aachen und die Arbeitsgruppe von Professor Theo Mora in Genua. Meine Aufenthalte in diesen Arbeitsgruppen haben meinen Blickwinkel f¨ur weit- ergehende Fragestellungen erweitert. An dieser Stelle m¨ochte ich mich bei allen jenen bedanken, die mich in dieser Zeit begleitet haben und so zum Entstehen und Gelingen dieser Arbeit beigetragen haben. Mein besonderer Dank gilt meinem akademischen Lehrer Professor Klaus Madlener, der meine akademische Ausbildung schon seit dem Grundstudium be- gleitet und meine Denk- und Arbeitsweise wesentlich gepr¨agt hat. Durch ihn habe ich gelernt, mich intensiv mit diesem Thema zu besch¨aftigen und mich dabei nie auf nur einen Blickwinkel zu beschr¨anken.
    [Show full text]
  • Programmable Digital Microcircuits - a Survey with Examples of Use
    - 237 - PROGRAMMABLE DIGITAL MICROCIRCUITS - A SURVEY WITH EXAMPLES OF USE C. Verkerk CERN, Geneva, Switzerland 1. Introduction For most readers the title of these lecture notes will evoke microprocessors. The fixed instruction set microprocessors are however not the only programmable digital mi• crocircuits and, although a number of pages will be dedicated to them, the aim of these notes is also to draw attention to other useful microcircuits. A complete survey of programmable circuits would fill several books and a selection had therefore to be made. The choice has rather been to treat a variety of devices than to give an in- depth treatment of a particular circuit. The selected devices have all found useful ap• plications in high-energy physics, or hold promise for future use. The microprocessor is very young : just over eleven years. An advertisement, an• nouncing a new era of integrated electronics, and which appeared in the November 15, 1971 issue of Electronics News, is generally considered its birth-certificate. The adver• tisement was for the Intel 4004 and its three support chips. The history leading to this announcement merits to be recalled. Intel, then a very young company, was working on the design of a chip-set for a high-performance calculator, for and in collaboration with a Japanese firm, Busicom. One of the Intel engineers found the Busicom design of 9 different chips too complicated and tried to find a more general and programmable solu• tion. His design, the 4004 microprocessor, was finally adapted by Busicom, and after further négociation, Intel acquired marketing rights for its new invention.
    [Show full text]
  • ABSTRACT Virtualizing Register Context
    ABSTRACT Virtualizing Register Context by David W. Oehmke Chair: Trevor N. Mudge A processor designer may wish for an implementation to support multiple reg- ister contexts for several reasons: to support multithreading, to reduce context switch overhead, or to reduce procedure call/return overhead by using register windows. Conventional designs require that each active context be present in its entirety, increasing the size of the register file. Unfortunately, larger register files are inherently slower to access and may lead to a slower cycle time or additional cycles of register access latency, either of which reduces overall performance. We seek to bypass the trade-off between multiple context support and register file size by mapping registers to memory, thereby decoupling the logical register requirements of active contexts from the contents of the physical register file. Just as caches and virtual memory allow a processor to give the illusion of numerous multi-gigabyte address spaces with an average access time approach- ing that of several kilobytes of SRAM, we propose an architecture that gives the illusion of numerous active contexts with an average access time approaching that of a single active context using a conventionally sized register file. This dis- sertation introduces the virtual context architecture, a new architecture that virtual- izes logical register contexts. Complete contexts, whether activation records or threads, are kept in memory and are no longer required to reside in their entirety in the physical register file. Instead, the physical register file is treated as a cache of the much larger memory-mapped logical register space. The implementation modifies the rename stage of the pipeline to trigger the movement of register val- ues between the physical register file and the data cache.
    [Show full text]
  • Recursion Theory Notes, Fall 2011 0.1 Introduction
    Recursion Theory Notes, Fall 2011 Lecturer: Lou van den Dries 0.1 Introduction Recursion theory (or: theory of computability) is a branch of mathematical logic studying the notion of computability from a rather theoretical point of view. This includes giving a lot of attention to what is not computable, or what is computable relative to any given, not necessarily computable, function. The subject is interesting on philosophical-scientific grounds because of the Church- Turing Thesis and its role in computer science, and because of the intriguing concept of Kolmogorov complexity. This course tries to keep touch with how recursion theory actually impacts mathematics and computer science. This impact is small, but it does exist. Accordingly, the first chapter of the course covers the basics: primitive recur- sion, partial recursive functions and the Church-Turing Thesis, arithmetization and the theorems of Kleene, the halting problem and Rice's theorem, recur- sively enumerable sets, selections and reductions, recursive inseparability, and index systems. (Turing machines are briefly discussed, but the arithmetization is based on a numerical coding of combinators.) The second chapter is devoted to the remarkable negative solution (but with positive aspects) of Hilbert's 10th Problem. This uses only the most basic notions of recursion theory, plus some elementary number theory that is worth knowing in any case; the coding aspects of natural numbers are combined here ingeniously with the arithmetic of the semiring of natural numbers. The last chapter is on Kolmogorov complexity, where concepts of recursion theory are used to define and study notions of randomness and information content.
    [Show full text]
  • Turing Machines [Fa’16]
    Models of Computation Lecture 6: Turing Machines [Fa’16] Think globally, act locally. — Attributed to Patrick Geddes (c.1915), among many others. We can only see a short distance ahead, but we can see plenty there that needs to be done. — Alan Turing, “Computing Machinery and Intelligence” (1950) Never worry about theory as long as the machinery does what it’s supposed to do. — Robert Anson Heinlein, Waldo & Magic, Inc. (1950) It is a sobering thought that when Mozart was my age, he had been dead for two years. — Tom Lehrer, introduction to “Alma”, That Was the Year That Was (1965) 6 Turing Machines In 1936, a few months before his 24th birthday, Alan Turing launched computer science as a modern intellectual discipline. In a single remarkable paper, Turing provided the following results: • A simple formal model of mechanical computation now known as Turing machines. • A description of a single universal machine that can be used to compute any function computable by any other Turing machine. • A proof that no Turing machine can solve the halting problem—Given the formal description of an arbitrary Turing machine M, does M halt or run forever? • A proof that no Turing machine can determine whether an arbitrary given proposition is provable from the axioms of first-order logic. This is Hilbert and Ackermann’s famous Entscheidungsproblem (“decision problem”). • Compelling arguments1 that his machines can execute arbitrary “calculation by finite means”. Although Turing did not know it at the time, he was not the first to prove that the Entschei- dungsproblem had no algorithmic solution.
    [Show full text]
  • Mathematical Society
    Notices of the American Mathematical Society November 1983, Issue 229 Volume 30, Number 7, Pages 713-840 Providence, Rhode Island USA ISSN 0002-9920 Calendar of AMS Meetings THIS CALENDAR lists all meetings which have been approved by the Council prior to the date this issue of the Notices was sent to press. The summer and annual meetings are joint meetings of the Mathematical Association of America and the Ameri· can Mathematical Society. The meeting dates which fall rather far in the future are subject to change; this is particularly true of meetings to which no numbers have yet been assigned. Programs of the meetings will appear in the issues indicated below. First and second announcements of the meetings will have appeared in earlier issues. ABSTRACTS OF PAPERS presented at a meeting of the Society are published in the journal Abstracts of papers presented to the American Mathematical Society in the issue corresponding to that of the Notices which contains the program of the meet­ ing. Abstracts should be submitted on special forms which are available in many departments of mathematics and from the office of the Society in Providence. Abstracts of papers to be presented at the meeting must be received at the headquarters of the Society in Providence, Rhode Island, on or before the deadline given below for the meeting. Note that the deadline for ab­ stracts submitted for consideration for presentation at special sessions is usually three weeks earlier than that specified below. For additional information consult the meeting announcement and the list of organizers of special sessions.
    [Show full text]
  • ECE 4750 Computer Architecture Topic 1: Microcoding
    ECE 4750 Computer Architecture Topic 1: Microcoding Christopher Batten School of Electrical and Computer Engineering Cornell University http://www.csl.cornell.edu/courses/ece4750 slide revision: 2013-09-01-10-42 Instruction Set Architecture Microcoded MIPS Processor Microcoding Discussion & Trends Agenda Instruction Set Architecture IBM 360 Instruction Set MIPS Instruction Set ISA to Microarchitecture Mapping Microcoded MIPS Processor Microcoded MIPS Microarchitecture #1 Microcoded MIPS Microarchitecture #2 Microcoding Discussion and Trends ECE 4750 T01: Microcoding 2 / 45 • Instruction Set Architecture • Microcoded MIPS Processor Microcoding Discussion & Trends Instruction Set Architecture I Contract between software & hardware Application Algorithm I Typically specified as all of the Programming Language programmer-visible state (registers & Operating System memory) plus the semantics of instructions Instruction Set Architecture Microarchitecture that operate on this state Register-Transfer Level IBM 360 was first line of machines to Gate Level I Circuits separate ISA from microarchitecture and Devices implementation Physics ... the structure of a computer that a machine language programmer must understand to write a correct (timing independent) program for that machine. — Amdahl, Blaauw, Brooks, 1964 ECE 4750 T01: Microcoding 3 / 45 • Instruction Set Architecture • Microcoded MIPS Processor Microcoding Discussion & Trends Compatibility Problem at IBM I By early 1960’s, IBM had several incompatible lines of computers! . Defense : 701 . Scientific : 704, 709, 7090, 7094 . Business : 702, 705, 7080 . Mid-Sized Business : 1400 . Decimal Architectures : 7070, 7072, 7074 I Each system had its own: . Instruction set . I/O system and secondary storage (tapes, drums, disks) . Assemblers, compilers, libraries, etc . Market niche ECE 4750 T01: Microcoding 4 / 45 • Instruction Set Architecture • Microcoded MIPS Processor Microcoding Discussion & Trends IBM 360: A General-Purpose Register Machine I Processor State .
    [Show full text]
  • Register Machines Storing Ordinals and Running for Ordinal Time
    2488 Register Machines Storing Ordinals and Running for Ordinal Time Ryan Siders 6/106 0 Outline • Assumptions about Motion and Limits on Programming • Without well-structured programming: • Well-Structured Programs: • Ordinal Computers with Well-Structured Programs: • Well-Structured programs halt. • Scratch Pad, Clocks, and Constants • Abstract Finite-time Computability Theory • Jacopini-Bohm Theorem holds if... • How many registers are universal? • 3 registers, No. 4 registers, Yes. • Main Theorem: Decide Any Constructible Set • Recursive Truth Predicate • Pop • Push • The Program to find Truth • Zero? 6/106 1 Assumptions about Motion and Limits on Programming Input, Memory, and Time for Hypercomputation: • Time is wellfounded so that the run is deterministic. • We want input and memory to be ordinals, too. • Then we can compare input length to run-time and define complexity classes. Definition An Ordinal Computer stores ordinals in its 0 < n < ! registers, runs for ordinal time, and executes a pro- gram that can alter the registers. At Limit times, we want the following to be determined: • which command line is active, and • what the values of the ordinals are. They are determined by: • assumptions on the behaviour of the registers and flow of control in the program (limiting the universe in which the programming model can exists), and • limitations on the class of programs that can be used. 6/106 2 Without well-structured programming: Definition: An Infinite-time Ordinal-storing Register Ma- chine is a register machine storing ordinal values and running for ordinal time, with a programming language including the three instructions: • Erase: Zero(x); • Increment: x + +; • Switch: if x = y goto i else j; whose state (the registers' values and active command) obeys the following three rules at a limit time: • If the command \Zero(x)" is called at each time t 2 T , then x is 0 at time sup T , too.
    [Show full text]
  • Jon Stokes Jon
    Inside the Machine the Inside A Look Inside the Silicon Heart of Modern Computing Architecture Computer and Microprocessors to Introduction Illustrated An Computers perform countless tasks ranging from the business critical to the recreational, but regardless of how differently they may look and behave, they’re all amazingly similar in basic function. Once you understand how the microprocessor—or central processing unit (CPU)— Includes discussion of: works, you’ll have a firm grasp of the fundamental concepts at the heart of all modern computing. • Parts of the computer and microprocessor • Programming fundamentals (arithmetic Inside the Machine, from the co-founder of the highly instructions, memory accesses, control respected Ars Technica website, explains how flow instructions, and data types) microprocessors operate—what they do and how • Intermediate and advanced microprocessor they do it. The book uses analogies, full-color concepts (branch prediction and speculative diagrams, and clear language to convey the ideas execution) that form the basis of modern computing. After • Intermediate and advanced computing discussing computers in the abstract, the book concepts (instruction set architectures, examines specific microprocessors from Intel, RISC and CISC, the memory hierarchy, and IBM, and Motorola, from the original models up encoding and decoding machine language through today’s leading processors. It contains the instructions) most comprehensive and up-to-date information • 64-bit computing vs. 32-bit computing available (online or in print) on Intel’s latest • Caching and performance processors: the Pentium M, Core, and Core 2 Duo. Inside the Machine also explains technology terms Inside the Machine is perfect for students of and concepts that readers often hear but may not science and engineering, IT and business fully understand, such as “pipelining,” “L1 cache,” professionals, and the growing community “main memory,” “superscalar processing,” and of hardware tinkerers who like to dig into the “out-of-order execution.” guts of their machines.
    [Show full text]