PC Processor Microarchitecture

Total Page:16

File Type:pdf, Size:1020Kb

PC Processor Microarchitecture 1 PC Processor Microarchitecture A Concise Review of the Techniques Used in Modern PC Processors by Keith Diefendorff For parallelism, today’s PC processors rely on pipelining and superscalar techniques to exploit instruction-level par- Having commandeered nearly all the performance- allelism (ILP). Pipelined processors overlap instructions in enhancing techniques used by their mainframe and super- time on common execution resources. Superscalar processors computer predecessors, the microprocessors in today’s PCs overlap instructions in space on separate resources. Both tech- employ a dizzying assemblage of microarchitectural features niques are used in combination. to achieve extraordinary levels of parallelism and speed. Unfortunately, performance gains from parallelism Enabled by astronomical transistor budgets, modern PC often fail to meet expectations. Although a four-stage pipe- processors are superscalar, deeply pipelined, out of order, and line, for example, overlaps the execution of four instructions, they even execute instructions speculatively. In this article, we as Figure 1 shows, it falls far short of a 4× performance boost. review the basic techniques used in these processors as well as The problem is pipeline stalls. Stalls arise from data hazards the tricks they employ to circumvent the two most challeng- (data dependencies), control hazards (changes in program ing performance obstacles: memory latency and branches. flow), and structural hazards (hardware resource conflicts), all of which sap pipeline efficiency. Two Paths to Performance Lengthening the pipeline, or superpipelining, divides The task normally assigned to chip architects is to design the instruction execution into more stages, each with a shorter highest-performance processor possible within a set of cost, cycle time; it does not, in general, shorten the execution time power, and size constraints established by market require- of instructions. In fact, it may increase execution time because ments. Within these constraints, application performance is stages rarely divide evenly and the frequency is set by the usually the best measure of success, although, sadly, the mar- longest stage. In addition, longer pipelines experience a ket often mistakes clock frequency for performance. higher percentage of stall cycles from hazards, thereby in- Two main avenues are open to designers trying to creasing the average cycles per instruction (CPI). Super- improve performance: making operations faster or executing scalar techniques suffer from similar inefficiencies. more of them in parallel. Operations can be made faster in The throughput gains from a longer pipeline, however, several ways. More advanced semiconductor processes make usually outweigh the CPI loss, so performance improves. But transistors switch faster and signals propagate faster. Using lengthening the pipeline has limits. As stages shrink, clock more transistors can reduce execution-unit latency (e.g., full skew and latch overheads (setup and hold times) consume a vs. partial multiplier arrays). Aggressive design methods can larger fraction of the cycle, leaving less usable time for logic. minimize the levels of logic needed to implement a given The challenge is to make the pipeline short enough for function (e.g., custom vs. standard-cell design) or to increase good efficiency but not so short that ILP and frequency are left circuit speed (e.g., dynamic vs. static circuits). lying on the table, i.e., an underpipelined condition. Today’s PC processors use pipelines of 5 to 12 stages. When making this decision, designers must keep in mind that frequency is often more important in the market than performance. Instr1 Fetch Issue Execute Write Instr2 Instr3 Prophetic Hardware for Long Pipelines Instr4 Branch prediction and speculative execution are tech- niques used to reduce pipeline stalls on control hazards. In a pipelined processor, conditional branches are often encoun- Instr1 tered before the data that will determine branch direction is Instr2 Instr3 ready. Because instructions are fetched ahead of execution, Instr4 correctly predicting unresolved branches allows the instruc- Instr5 Instr6 tion fetcher to keep the instruction queue filled with instruc- Instr7 tions that have a high probability of being used. Instr8 Instr9 Some processors take the next step, actually executing instructions speculatively past unresolved conditional Figure 1. Pipelines overlap the execution of instructions in time. Lengthening the pipeline increases the number of instructions exe- branches. This technique avoids the control-hazard stall cuted in a given time period. Longer pipelines, however, suffer altogether when the branch goes in the predicted direction. from a higher percentage of stalls (not shown). On mispredictions, however, the pipeline must be flushed, ©MICRODESIGN RESOURCES JULY 12, 1999 MICROPROCESSOR REPORT 2 instruction fetch redirected, and the pipeline refilled. Statis- Although compilers can statically reschedule instruc- tically, prediction and speculation dramatically reduce stalls. tions, they are hampered by incomplete knowledge of run- How dramatically depends on prediction accuracy. time information. Load-use penalties, for example, are Branch predictors range in sophistication from simple resistant to static rescheduling because their length is gener- static predictors (compiler or heuristic driven), which ally unpredictable at compile time. It is simply impossible to achieve 65–85% accuracy, to complex dynamic predictors find enough independent instructions to cover the worst- that can achieve 98% accuracy or more. Since one in five case number of load-delay slots in every load. instructions is typically a conditional branch, high accuracy is Static rescheduling is also constrained by register name- essential, especially for machines with long pipelines and, space and by ambiguous dependencies between memory therefore, with large mispredict penalties. As a result, most instructions. A large register namespace is required for good modern processors employ dynamic predictors. register allocation, for freedom in rearranging instructions, and for loop unrolling. Register limitations are especially The Past Predicts the Future severe in x86 processors, which have only eight general- The simplest dynamic predictor is the branch history table purpose registers. In-order processors—which issue, execute, (BHT), a small cache indexed by the address of the branch complete, and retire instructions in strict program order— being predicted. Simple BHTs record one-bit histories of the must rely entirely on static rescheduling and can suffer a large direction each branch took the last time it executed. More number of pipeline stalls. sophisticated BHTs use two-bit histories, which add hystere- Therefore, most current PC processors implement sis to improve prediction accuracy on loop branches. Even dynamic instruction rescheduling to some degree. The sim- more sophisticated schemes use two-level predictors with plest out-of-order processors issue instructions in order but longer per-branch histories that index into pattern tables allow them to execute and complete out of order. Processors containing two-bit predictors (see MPR 3/27/95, p. 17). of this type use register scoreboarding to interlock the pipe- A simplified version of the two-level predictor uses a line, stalling instruction issue when an instruction’s operands single global-history register of recent branch directions to aren’t ready. Such processors can achieve somewhat more index into the BHT. The GShare enhancement (see MPR parallelism than in-order processors by permitting instruc- 11/17/97, p. 22) adds per-branch sensitivity by hashing a few tions to execute in parallel through execution units with dif- bits of the branch address with the global-history register, as ferent or variable latencies. Figure 2 shows. The agrees-mode enhancement encodes the Even simple out-of-order processors require complex prediction as agreement or disagreement with a static pre- hardware to reorder results before the corresponding in- diction, thereby avoiding excessive mispredictions when structions are retired (removed from the machine). Although multiple active branches map to the same BHT entry. In strict result ordering is not needed from a data-flow perspec- architectures with no static-prediction opcode bits, such as tive, it is required to maintain precise exceptions (the appear- the x86, the static prediction must be based on branch ance of in-order execution following an interrupt) and to heuristics (e.g., backward: predict taken). recover from mispredicted speculative execution. Some processors predict the target instruction stream as The most common reordering method is the reorder well as the direction. Target predictions are made with a buffer (ROB), which buffers results until they can be written branch target address cache (BTAC), which caches the to the register file in program order. Accessing operands address to which control was transferred the last time the from the reorder buffer, which is needed for reasonable per- branch was taken. BTACs are sometimes combined with the formance, requires an associative lookup to locate the most BHT into a branch target buffer (BTB). Instead of a BTAC, recent version of the operand. some processors use a branch target instruction cache (BTIC), which caches the first few instructions down the tar- Global Branch History 01 get path so the pipeline can be primed without an inline fetch Register 11 Branch
Recommended publications
  • REPORT Compaq Chooses SMT for Alpha Simultaneous Multithreading
    VOLUME 13, NUMBER 16 DECEMBER 6, 1999 MICROPROCESSOR REPORT THE INSIDERS’ GUIDE TO MICROPROCESSOR HARDWARE Compaq Chooses SMT for Alpha Simultaneous Multithreading Exploits Instruction- and Thread-Level Parallelism by Keith Diefendorff Given a full complement of on-chip memory, increas- ing the clock frequency will increase the performance of the As it climbs rapidly past the 100-million- core. One way to increase frequency is to deepen the pipeline. transistor-per-chip mark, the micro- But with pipelines already reaching upwards of 12–14 stages, processor industry is struggling with the mounting inefficiencies may close this avenue, limiting future question of how to get proportionally more performance out frequency improvements to those that can be attained from of these new transistors. Speaking at the recent Microproces- semiconductor-circuit speedup. Unfortunately this speedup, sor Forum, Joel Emer, a Principal Member of the Technical roughly 20% per year, is well below that required to attain the Staff in Compaq’s Alpha Development Group, described his historical 60% per year performance increase. To prevent company’s approach: simultaneous multithreading, or SMT. bursting this bubble, the only real alternative left is to exploit Emer’s interest in SMT was inspired by the work of more and more parallelism. Dean Tullsen, who described the technique in 1995 while at Indeed, the pursuit of parallelism occupies the energy the University of Washington. Since that time, Emer has of many processor architects today. There are basically two been studying SMT along with other researchers at Washing- theories: one is that instruction-level parallelism (ILP) is ton. Once convinced of its value, he began evangelizing SMT abundant and remains a viable resource waiting to be tapped; within Compaq.
    [Show full text]
  • Prozessorarchitektur Am Beispiel Des Amdathlon
    PROZESSORARCHITEKTUR AM BEISPIEL DES AMD ATHLON AUSGEARBEITET VON ALEXANDER TABAKOFF Betreuender Lehrer: Prof. Wolfgang Schinwald VERÖFFENTLICHT AM 26.2.2001 PROZESSORARCHITEKTUR INHALTSVERZEICHNIS: 1 Historische / allgemeine Einführung 1.1Die Anwendungsbereiche von Prozessoren 1.2Der erste Prozessor 1.3Die Entwicklung bis zum 586 1.4Der AMD Athlon und der Pentium III - Entwicklungsgeschichte 2 Grundlegende Dinge zur Prozessorarchitektur und dem Bau von Prozessoren 2.1Physikalisch 2.1.1Der Aufbau eines Transistors 2.1.2Die Auswirkungen in die Praxis 2.2Logisch 2.3Die Herstellung von Prozessoren und ihre Grenzen 2.4Der Von-Neumann-Rechner 3 Die Prozessorarchitektur des AMD Athlon im Vergleich zu seinen Konkurrenten 3.1Das Design des AMD Athlon 3.2Das Bussytem des AMD Athlon 3.3Die Cachearchitektur des AMD Athlon 3.4Vor- und Nachteile gegenüber anderen Designs 3.5Interview mit Jan Gütter, Public Relations Sprecher von AMD 4 Anhang 4.1Der Grund dieser Arbeit 4.2Glossar 4.3Literaturverzeichnis 4.4Begleitprotokoll 4.5Bildnachweis Inhaltsverzeichnis: - Seite 2 PROZESSORARCHITEKTUR 1 HISTORISCHE / ALLGEMEINE EINFÜHRUNG 1.1Die Anwendungsbereiche von Prozessoren Prozessoren haben heute verschiedenste Anwendungsbereiche. Sie werden in Autos, Set Top Boxen, Spielekonsolen, Handys, Taschenrechnern, PCs usw. verwendet. Dabei macht der Marktanteil der PC Prozessoren nur rund 2%1 aus. Trotz dieser vergleichsweise geringen Produktion genießen PC Prozessoren einen bedeutend höheren Bekanntheitsgrad. Fast jeder kennt PC Prozessoren wie den Intel Pentium
    [Show full text]
  • A Bibliography of Publications in IEEE Micro
    A Bibliography of Publications in IEEE Micro Nelson H. F. Beebe University of Utah Department of Mathematics, 110 LCB 155 S 1400 E RM 233 Salt Lake City, UT 84112-0090 USA Tel: +1 801 581 5254 FAX: +1 801 581 4148 E-mail: [email protected], [email protected], [email protected] (Internet) WWW URL: http://www.math.utah.edu/~beebe/ 16 September 2021 Version 2.108 Title word cross-reference -Core [MAT+18]. -Cubes [YW94]. -D [ASX19, BWMS19, DDG+19, Joh19c, PZB+19, ZSS+19]. -nm [ABG+16, KBN16, TKI+14]. #1 [Kah93i]. 0.18-Micron [HBd+99]. 0.9-micron + [Ano02d]. 000-fps [KII09]. 000-Processor $1 [Ano17-58, Ano17-59]. 12 [MAT 18]. 16 + + [ABG+16]. 2 [DTH+95]. 21=2 [Ste00a]. 28 [BSP 17]. 024-Core [JJK 11]. [KBN16]. 3 [ASX19, Alt14e, Ano96o, + AOYS95, BWMS19, CMAS11, DDG+19, 1 [Ano98s, BH15, Bre10, PFC 02a, Ste02a, + + Ste14a]. 1-GHz [Ano98s]. 1-terabits DFG 13, Joh19c, LXB07, LX10, MKT 13, + MAS+07, PMM15, PZB+19, SYW+14, [MIM 97]. 10 [Loc03]. 10-Gigabit SCSR93, VPV12, WLF+08, ZSS+19]. 60 [Gad07, HcF04]. 100 [TKI+14]. < [BMM15]. > [BMM15]. 2 [Kir84a, Pat84, PSW91, YSMH91, ZACM14]. [WHCK18]. 3 [KBW95]. II [BAH+05]. ∆ 100-Mops [PSW91]. 1000 [ES84]. 11- + [Lyl04]. 11/780 [Abr83]. 115 [JBF94]. [MKG 20]. k [Eng00j]. µ + [AT93, Dia95c, TS95]. N [YW94]. x 11FO4 [ASD 05]. 12 [And82a]. [DTB01, Dur96, SS05]. 12-DSP [Dur96]. 1284 [Dia94b]. 1284-1994 [Dia94b]. 13 * [CCD+82]. [KW02]. 1394 [SB00]. 1394-1955 [Dia96d]. 1 2 14 [WD03]. 15 [FD04]. 15-Billion-Dollar [KR19a].
    [Show full text]
  • Optimizing SIMD Execution in HW/SW Co-Designed Processors
    Optimizing SIMD Execution in HW/SW Co-designed Processors Rakesh Kumar Department of Computer Architecture Universitat Politècnica de Catalunya Advisors: Alejandro Martínez Intel Barcelona Research Center Antonio González Intel Barcelona Research Center Universitat Politècnica de Catalunya A thesis submitted in fulfillment of the requirements for the degree of Doctor of Philosophy / Doctor per la UPC ABSTRACT SIMD accelerators are ubiquitous in microprocessors from different computing domains. Their high compute power and hardware simplicity improve overall performance in an energy efficient manner. Moreover, their replicated functional units and simple control mechanism make them amenable to scaling to higher vector lengths. However, code generation for these accelerators has been a challenge from the days of their inception. Compilers generate vector code conservatively to ensure correctness. As a result they lose significant vectorization opportunities and fail to extract maximum benefits out of SIMD accelerators. This thesis proposes to vectorize the program binary at runtime in a speculative manner, in addition to the compile time static vectorization. There are different environments that support runtime profiling and optimization support required for dynamic vectorization, one of most prominent ones being: 1) Dynamic Binary Translators and Optimizers (DBTO) and 2) Hardware/Software (HW/SW) Co-designed Processors. HW/SW co-designed environment provides several advantages over DBTOs like transparent incorporations of new hardware features, binary compatibility, etc. Therefore, we use HW/SW co-designed environment to assess the potential of speculative dynamic vectorization. Furthermore, we analyze vector code generation for wider vector units and find out that even though SIMD accelerators are amenable to scaling from hardware point of view, vector code generation at higher vector length is even more challenging.
    [Show full text]
  • UNIVERSITY of CALIFORNIA, SAN DIEGO Holistic Design for Multi-Core Architectures a Dissertation Submitted in Partial Satisfactio
    UNIVERSITY OF CALIFORNIA, SAN DIEGO Holistic Design for Multi-core Architectures A dissertation submitted in partial satisfaction of the requirements for the degree Doctor of Philosophy in Computer Science (Computer Engineering) by Rakesh Kumar Committee in charge: Professor Dean Tullsen, Chair Professor Brad Calder Professor Fred Chong Professor Rajesh Gupta Dr. Norman P. Jouppi Professor Andrew Kahng 2006 Copyright Rakesh Kumar, 2006 All rights reserved. The dissertation of Rakesh Kumar is approved, and it is acceptable in quality and form for publication on micro- film: Chair University of California, San Diego 2006 iii DEDICATIONS This dissertation is dedicated to friends, family, labmates, and mentors { the ones who taught me, indulged me, loved me, challenged me, and laughed with me, while I was also busy working on my thesis. To Professor Dean Tullsen for teaching me the values of humility, kind- ness,and caring while trying to teach me football and computer architecture. For always encouraging me to do the right thing. For always letting me be myself. For always believing in me. For always challenging me to dream big. For all his wisdom. And for being an adviser in the truest sense, and more. To Professor Brad Calder. For always caring about me. For being an inspiration. For his trust. For making me believe in myself. For his lies about me getting better at system administration and foosball even though I never did. To Dr Partha Ranganathan. For always being there for me when I would get down on myself. And that happened often. For the long discussions on life, work, and happiness.
    [Show full text]
  • Sony's Emotionally Charged Chip
    VOLUME 13, NUMBER 5 APRIL 19, 1999 MICROPROCESSOR REPORT THE INSIDERS’ GUIDE TO MICROPROCESSOR HARDWARE Sony’s Emotionally Charged Chip Killer Floating-Point “Emotion Engine” To Power PlayStation 2000 by Keith Diefendorff rate of two million units per month, making it the most suc- cessful single product (in units) Sony has ever built. While Intel and the PC industry stumble around in Although SCE has cornered more than 60% of the search of some need for the processing power they already $6 billion game-console market, it was beginning to feel the have, Sony has been busy trying to figure out how to get more heat from Sega’s Dreamcast (see MPR 6/1/98, p. 8), which has of it—lots more. The company has apparently succeeded: at sold over a million units since its debut last November. With the recent International Solid-State Circuits Conference (see a 200-MHz Hitachi SH-4 and NEC’s PowerVR graphics chip, MPR 4/19/99, p. 20), Sony Computer Entertainment (SCE) Dreamcast delivers 3 to 10 times as many 3D polygons as and Toshiba described a multimedia processor that will be the PlayStation’s 34-MHz MIPS processor (see MPR 7/11/94, heart of the next-generation PlayStation, which—lacking an p. 9). To maintain king-of-the-mountain status, SCE had to official name—we refer to as PlayStation 2000, or PSX2. do something spectacular. And it has: the PSX2 will deliver Called the Emotion Engine (EE), the new chip upsets more than 10 times the polygon throughput of Dreamcast, the traditional notion of a game processor.
    [Show full text]
  • Commemorative Booklet for the Thirty-Fifth Asilomar Microcomputer Workshop April 15-17, 2009
    35 Commemorative Booklet for the Thirty-Fifth Asilomar Microcomputer Workshop April 15-17, 2009 Programs from the 1975-2009 Workshops This file available at www.amw.org AMW: 3dh Workshop Prologue - Ted Laliotis The Asilomar Microcomputer Workshop (AMW) has played a very important role during its 30 years ofexistence. Perhaps, that is why it continues to be well attended. The workshop was founded in 1975 as an IEEE technical workshop sponsored by the Western Area Committee ofthe IEEE Computer Society. The intentional lack of written proceedings and the exclusion of general press representatives was perhaps the most distinctive characteristic of AMW that made it so special and successful. This encouraged the scientists and engineers who were at the cutting edge ofthe technology, the movers and shakers that shaped Silicon Valley, the designers of the next generation microprocessors, to discuss and debate freely the various issues facing microprocessors. In fact, many features, or lack of, were born during the discussions and debates at AMW. We often referred to AMW and its attendees as the bowels of Silicon Valley, even though attendees came from all over the country, and the world. Another characteristic that made AMW special was the "required" participation and contribution by all attendees. Every applicant to attend AMW had to convince the committee that he had something to contribute by speaking during one of the sessions or during the open mike session. In the event that someone slipped through and was there only to listen, that person was not invited back the following year. The decades ofthe 70's and 80's were probably the defining decades for the amazing explosion of microcomputers.
    [Show full text]
  • Selective Vectorization for Short-Vector Instructions Samuel Larsen, Rodric Rabbah, and Saman Amarasinghe
    Computer Science and Artificial Intelligence Laboratory Technical Report MIT-CSAIL-TR-2009-064 December 18, 2009 Selective Vectorization for Short-Vector Instructions Samuel Larsen, Rodric Rabbah, and Saman Amarasinghe massachusetts institute of technology, cambridge, ma 02139 usa — www.csail.mit.edu Selective Vectorization for Short-Vector Instructions Samuel Larsen∗ Rodric Rabbah∗ Saman Amarasinghe VMware IBM Research MIT CSAIL Abstract Multimedia extensions are nearly ubiquitous in today’s general-purpose processors. These extensions consist primarily of a set of short-vector instructions that apply the same opcode to a vector of operands. Vector instructions introduce a data-parallel component to processors that exploit instruction-level par- allelism, and present an opportunity for increased performance. In fact, ignoring a processor’s vector opcodes can leave a significant portion of the available resources unused. In order for software developers to find short-vector instructions generally useful, however, the compiler must target these extensions with complete transparency and consistent performance. This paper describes selective vectorization, a technique for balancing computation across a proces- sor’s scalar and vector units. Current approaches for targeting short-vector instructions directly adopt vectorizing technology first developed for supercomputers. Traditional vectorization, however, can lead to a performance degradation since it fails to account for a processor’s scalar resources. We formulate selective vectorization in the context of software pipelining. Our approach creates software pipelines with shorter initiation intervals, and therefore, higher performance. A key aspect of selective vectorization is its ability to manage transfer of operands between vector and scalar instructions. Even when operand transfer is expensive, our technique is sufficiently sophisticated to achieve significant performance gains.
    [Show full text]
  • HP, Intel Complete IA-64 Rollout:4/10/00
    MICROPROCESSOR www.MPRonline.com THE REPORTINSIDER’S GUIDE TO MICROPROCESSOR HARDWARE HP, INTEL COMPLETE IA-64 ROLLOUT Virtual Memory, Interrupts More Conventional Than ISA By Keith Diefendorff {4/10/00-01} HP and Intel have finally finished the long-drawn-out task of revealing the IA-64 architecture, a process that began more than two years ago at the 1997 Microprocessor Forum. Speaking at Intel’s spring developer forum (IDF), IA-64-architects Jerry Huck and Rumi Zahir laid out the details of the IA-64 system architecture, including descrip- implemented by Windows NT, Linux, BSD, VMS, and Mach- tions of its virtual-memory mechanisms, memory protection, based versions of Unix—or the single-address-space (SAS) multiprocessor coherence, and interrupt structure. model used in proprietary versions of Unix, such as HP-UX Unlike the unconventional approach HP and Intel took and IBM’s AIX. (The SAS model is also referred to as the with IA-64’s VLIW instruction-set architecture (ISA) (see global-address-space model.) Historically, processors have MPR 5/31/99-01, “IA-64: A Parallel Instruction Set”), the supported one model or the other, making them more or less companies took a far less radical tack on system architecture. adaptable to a given OS. Borrowing liberally from existing RISC processors, especially MAS operating systems typically use unique process from PA-RISC and PowerPC, IA-64 processors will require identifiers (PIDs) and simple multilevel-indexed page-table little redesign of operating-system internals. Although the vendors released system-architecture 64-bit Logical Address details to selected IA-64 software developers under nondis- Page Offset (12–28 bits) closure agreements some time ago, the recent IDF disclosure 63 61 allows any company to create any level of software, right Virtual Virtual down to and including operating-system kernels and plat- Region Page Number Number Virtual Regions form BIOSs.
    [Show full text]
  • Low Overhead Memory Subsystem Design for a Multicore Parallel DSP Processor
    Linköping Studies in Science and Technology Dissertation No. 1532 Low Overhead Memory Subsystem Design for a Multicore Parallel DSP Processor Jian Wang Department of Electrical Engineering Linköping University SE-581 83 Linköping, Sweden Linköping 2014 ISBN 978-91-7519-556-8 ISSN 0345-7524 ii Low Overhead Memory Subsystem Design for a Multicore Parallel DSP Processor Jian Wang ISBN 978-91-7519-556-8 Copyright ⃝c Jian Wang, 2014 Linköping Studies in Science and Technology Dissertation No. 1532 ISSN 0345-7524 Department of Electrical Engineering Linköping University SE-581 83 Linköping Sweden Phone: +46 13 28 10 00 Author e-mail: [email protected] Cover image Combined Star and Ring onchip interconnection of the ePUMA multicore DSP. Parts of this thesis are reprinted with permission from the IEEE. Printed by UniTryck, Linköping University Linköping, Sweden, 2014 Abstract The physical scaling following Moore’s law is saturated while the re- quirement on computing keeps growing. The gain from improving sili- con technology is only the shrinking of the silicon area, and the speed- power scaling has almost stopped in the last two years. It calls for new parallel computing architectures and new parallel programming meth- ods. Traditional ASIC (Application Specific Integrated Circuits) hardware has been used for acceleration of Digital Signal Processing (DSP) subsys- tems on SoC (System-on-Chip). Embedded systems become more com- plicated, and more functions, more applications, and more features must be integrated in one ASIC chip to follow up the market requirements. At the same time, the product lifetime of a SoC with ASIC has been much reduced because of the dynamic market.
    [Show full text]
  • Commemorative Booklet for the Thirty-Fifth Asilomar Microcomputer Workshop April 15-17, 2009 Programs from the 1975-2009 Worksho
    35 Commemorative Booklet for the Thirty-Fifth Asilomar Microcomputer Workshop April 15-17, 2009 Programs from the 1975-2009 Workshops This file available at www.amw.org AMW: 3dh Workshop Prologue - Ted Laliotis The Asilomar Microcomputer Workshop (AMW) has played a very important role during its 30 years ofexistence. Perhaps, that is why it continues to be well attended. The workshop was founded in 1975 as an IEEE technical workshop sponsored by the Western Area Committee ofthe IEEE Computer Society. The intentional lack of written proceedings and the exclusion of general press representatives was perhaps the most distinctive characteristic of AMW that made it so special and successful. This encouraged the scientists and engineers who were at the cutting edge ofthe technology, the movers and shakers that shaped Silicon Valley, the designers of the next generation microprocessors, to discuss and debate freely the various issues facing microprocessors. In fact, many features, or lack of, were born during the discussions and debates at AMW. We often referred to AMW and its attendees as the bowels of Silicon Valley, even though attendees came from all over the country, and the world. Another characteristic that made AMW special was the "required" participation and contribution by all attendees. Every applicant to attend AMW had to convince the committee that he had something to contribute by speaking during one of the sessions or during the open mike session. In the event that someone slipped through and was there only to listen, that person was not invited back the following year. The decades ofthe 70's and 80's were probably the defining decades for the amazing explosion of microcomputers.
    [Show full text]
  • Altivec Extension to Powerpc Accelerates Media Processing
    ALTIVEC EXTENSION TO POWERPC ACCELERATES MEDIA PROCESSING DESIGNED AROUND THE PREMISE THAT MULTIMEDIA WILL BE THE PRIMARY CONSUMER OF PROCESSING CYCLES IN FUTURE PCS, ALTIVEC—WHICH APPLE CALLS THE VELOCITY ENGINE—INCREASES PERFORMANCE ACROSS A BROAD SPECTRUM OF MEDIA PROCESSING APPLICATIONS. There is a clear trend in personal com- extension to a general-purpose architecture. puting toward multimedia-rich applications. But the similarity ends there. Whereas the These applications will incorporate a wide vari- other extensions were obviously constrained ety of multimedia technologies, including audio by backward compatibility and a desire to and video compression, 2D image processing, limit silicon investment to a small fraction of 3D graphics, speech and handwriting recogni- the processor die area, the primary goal for Keith Diefendorff tion, media mining, and narrow-/broadband AltiVec was high functionality. It was designed signal processing for communication. from scratch around the premise that multi- Microprocessor Report In response to this demand, major micro- media will become the primary consumer of processor vendors have announced architec- processing cycles8 in future PCs and therefore tural extensions to their general-purpose deserves first-class treatment in the CPU. Pradeep K. Dubey processors in an effort to improve their multi- Unlike most other extensions, which over- media performance. Intel extended IA-32 with load their floating-point (FP) registers to IBM Research Division MMX1 and SSE (alias KNI),2 Sun enhanced accommodate multimedia data, AltiVec ded- Sparc with VIS,3 Hewlett-Packard added icates a large new register file exclusively to it. MAX4 to its PA-RISC architecture, Silicon Although overloading the FP registers avoids Ron Hochsprung Graphics extended the MIPS architecture with new architectural state, eliminating the need MDMX,5 and Digital (now Compaq) added to modify the operating system, it also signif- Apple Computer MVI to Alpha.
    [Show full text]