Multicore: Why Is It Happening Now? Eller Hur Mår Moore’S Lag?

Total Page:16

File Type:pdf, Size:1020Kb

Multicore: Why Is It Happening Now? Eller Hur Mår Moore’S Lag? Multicore: Why is it happening now? eller Hur Mår Moore’s Lag? Erik Hagersten Uppsala Universitet Are we hitting the wall now? Pop: Can the transistors be made even smaller and faster? Performance Possible path, [log] but requires a paradigm shift 1000 ? Business as 100 ” w usual… La s re o Well uhm … 10 o M “ The transistors can p: Po be made smaller and PDC 1 Summer faster, but there are School other problems / 2010 Institutionen för informationsteknologi | www.it.uu.seNowMC 2 ©Erik Hagersten| user.it.uu.se/~ehYear Darling, I shrunk the computer Mainframes Super Minis: Microprocessor: Mem Multicore: Many CPUs on a Mem PDC Summer chip! School 2010 Institutionen för informationsteknologi | www.it.uu.se MC 3 ©Erik Hagersten| user.it.uu.se/~eh Multi-core CPUs: Ageia PhysX, a multi-core physics processing unit. Ambric Am2045, a 336-core Massively Parallel Processor Array (MPPA) AMD Athlon 64, Athlon 64 FX and Athlon 64 X2 family, dual-core desktop processors. Opteron, dual- and quad-core server/workstation processors. Phenom, triple- and quad-core desktop processors. Sempron X2, dual-core entry level processors. Turion 64 X2, dual-core laptop processors. Radeon and FireStream multi-core GPU/GPGPU (10 cores, 16 5-issue wide superscalar stream processors per core) ARM MPCore is a fully synthesizable multicore container for ARM9 and ARM11 processor cores, intended for high-performance embedded and entertainment applications. Azul Systems Vega 2, a 48-core processor. Broadcom SiByte SB1250, SB1255 and SB1455. Cradle Technologies CT3400 and CT3600, both multi-core DSPs. Cavium Networks Octeon, a 16-core MIPS MPU. HP PA-8800 and PA-8900, dual core PA-RISC processors. IBM POWER4, the world's first dual-core processor, released in 2001. POWER5, a dual-core processor, released in 2004. POWER6, a dual-core processor, released in 2007. PowerPC 970MP, a dual-core processor, used in the Apple Power Mac G5. Xenon, a triple-core, SMT-capable, PowerPC microprocessor used in the Microsoft Xbox 360 game console. IBM, Sony, and Toshiba Cell processor, a nine-core processor with one general purpose PowerPC core and eight specialized SPUs (Synergystic Processing Unit) optimized for vector operations used in the Sony PlayStation 3. Infineon Danube, a dual-core, MIPS-based, home gateway processor. Intel Celeron Dual Core, the first dual-core processor for the budget/entry-level market. Core Duo, a dual-core processor. Core 2 Duo, a dual-core processor. Core 2 Quad, a quad-core processor. Core i7, a quad-core processor, the successor of the Core 2 Duo and the Core 2 Quad. Itanium 2, a dual-core processor. Pentium D, a dual-core processor. Teraflops Research Chip (Polaris), an 3.16 GHz, 80-core processor prototype, which the company says will be released within the next five years[6]. Xeon dual-, quad- and hexa-core processors. IntellaSys seaForth24, a 24-core processor. Nvidia GeForce 9 multi-core GPU (8 cores, 16 scalar stream processors per core) GeForce 200 multi-core GPU (10 cores, 24 scalar stream processors per core) Tesla multi-core GPGPU (8 cores, 16 scalar stream processors per core) Parallax Propeller P8X32, an eight-core microcontroller. picoChip PC200 series 200-300 cores per device for DSP & wireless Rapport Kilocore KC256, a 257-core microcontroller with a PowerPC core and 256 8-bit "processing elements". Raza Microelectronics XLR, an eight-core MIPS MPU Sun Microsystems PDC UltraSPARC IV and UltraSPARC IV+, dual-core processors. UltraSPARC T1, an eight-core, 32-thread processor. Summer UltraSPARC T2, an eight-core, 64-concurrent-thread processor. Texas Instruments TMS320C80 MVP, a five-core multimedia[source: video processor. Wikipedia] School Tilera TILE64, a 64-core processor XMOS Software Defined Silicon quad-core XS1-G4 2010 Institutionen för informationsteknologi | www.it.uu.se MC 4 ©Erik Hagersten| user.it.uu.se/~eh High-performance single-core CPUs: PDC Summer School 2010 Institutionen för informationsteknologi | www.it.uu.se MC 5 ©Erik Hagersten| user.it.uu.se/~eh Many shapes and forms... Mem I/F Mem FSB Mem m m m m € I/F I/F € € C C C C $ $ $ $ C C C C $ ¢ ¢ ¢ ¢ ¢ ¢ ¢ ¢ m m m m ¢ C C C C C C C C C AMD Intel Core2 IBM Cell PDC Summer School 2010 Institutionen för informationsteknologi | www.it.uu.se MC 6 ©Erik Hagersten| user.it.uu.se/~eh Outline Why multicore now? Technology reasons Performance bottlenecks in MCs Commercial offerings Reflection for the future PDC Summer School 2010 Institutionen för informationsteknologi | www.it.uu.se MC 7 ©Erik Hagersten| user.it.uu.se/~eh Everybody is doing is! But, why now? Perf Multi core [log] Single core time 2007 1. Not enough ILP to get payoff from using more transistors 2. Signal propagation delay » transistor delay 3. 2 PDC Power consumption Pdyn ~ C • f • V Summer School 2010 Institutionen för informationsteknologi | www.it.uu.se MC 8 ©Erik Hagersten| user.it.uu.se/~eh Microprocessors today: Whatever it takes to run Exploring ILP (instruction-level parallelism): Faster clocks Superscalar Pipelines Branch Prediction one Prefetching Out-of-Order ExecutionÎ Trace Cache Deep pipelines program fast. Speculation Predicate Execution Advanced Load Address Table PDC Return Address Stack Summer School ….Bad News #1: 2010 We have already explored most ILP Institutionen för informationsteknologi(ins truction-level parallelism) | www.it.uu.se MC 9 ©Erik Hagersten | user.it.uu.se/~eh Bad News #2 Looong#2: Future wire delay Technology Îslow CPUs cycle time 830 ps (1.2GHz) 170ps Latency for a 5mm trace (RC model) 390ps PDC Summer School Quantitative data 2000and trends 2005 according 2010 2015 to V. Agarwal et 70al., ps ISCA(14 GHz) 2000 2010 Based on SIA Institutionen för informationsteknologi(Semiconductor Industry Assoc | www.it.uu.se MC 10 iation) prediction, 1999 År ©Erik Hagersten | user.it.uu.se/~eh #2:Bad How News much of#2 the chip area can youLooong reach wire in one delay cycle? Îslow CPUs Span -- Fraction of chip reachable in one cycle 1.0 100% 0.75 0.50 SIA 0.25 0.4 - 1.4 % 2000 2005 2010 2015 Year PDC .25 .18 .13 .10 .070 .050 .0035 Feature size(micron) Summer School 2010 Institutionen för informationsteknologi | www.it.uu.se MC 11 ©Erik Hagersten| user.it.uu.se/~eh Bad News #2:wire delay Î Multiple small CPUs with private L1$ CPUCPUCPUCPU L1L1L1 L1L1 cache cachecache Externa L1 cache L3 L3 L2 Cache Externa Ctrl L3 PDC Summer School 2010 Institutionen för informationsteknologi | www.it.uu.se MC 12 ©Erik Hagersten| user.it.uu.se/~eh Bad News #3: Power consumption Î Lower frequency Î lower voltage 2 2 Pdyn = C * f * V ≈ area * freq * voltage CPU CPU CPU freq=f vs. freq=f/2 freq=f/2 2 2 Pdyn(C, f, V) = CfV Pdyn(2C, f/2, <V) < CfV CPU CPU CPU vs. freq=f CPU CPU freq = f/2 2 2 Pdyn(C, f, V) = CfV Pdyn(C, f/2, <V) < ½ CfV PDC Summer School 2010 Institutionen för informationsteknologi | www.it.uu.se MC 13 ©Erik Hagersten| user.it.uu.se/~eh Example: Freq. Scaling speed 0.87 1.0 pow 0.51 freq speed pow freq speed freq speed 1.20 1.13 1.51 0.80 0.87 pow 0.80 0.87 pow 0.51 0.51 20% higher freq. 20% lower freq. 20% lower freq. Two cores PDC Summer School 2010 Institutionen för informationsteknologi | www.it.uu.se MC 14 ©Erik Hagersten| user.it.uu.se/~eh Darling, I shrunk the computer Sequential execution ( Mainframes ≈ Super Minis: one program) Microprocessor: Mem Paradigm Shift Parallel Chip Multiprocessor (CMP): Mem Apps (TLP) PDC Summer A multiprocessor on a chip! School 2010 Institutionen för informationsteknologi | www.it.uu.se MC 15 ©Erik Hagersten| user.it.uu.se/~eh How to create threads? e e c c n n ie ie c c Seq. Seq. S S t t Prog. eProg. e Not ideal for k k HARD!! c c o o multi cores R R Parallel Parallelizing Modeling Prog1 Prog2 OS Prog compiler Thread1 Thread2 Thread3 Thread1 Thread2 Thread3 Throughput Computing Capability Computing •“Multitasking” •Weather simulation PDC •Virtualization •Computer games Summer •Concurrency •… School •… 2010 Institutionen för informationsteknologi | www.it.uu.se MC 16 ©Erik Hagersten| user.it.uu.se/~eh Shared Bottlenecks CPU CPU CPU CPU L1 L1 L1 L1 L2 Cache L1 L1 L1 L1 Bandwidth CPU CPU CPU CPU • Cache sharing effects • Cache utilization • Loop order • Data allocation • Data usage… PDC • Data reuse Shared Summer • Tiling (aka Blocking) Resources School • Fusion… 2010 Institutionen för informationsteknologi | www.it.uu.se MC 17 ©Erik Hagersten| user.it.uu.se/~eh Example: Poor Throughput Scaling! Example: 470.LBM "Lattice Boltzmann Method" to simulate incompressible fluids in 3D 3,5 3 2,5 2 1,5 1.0 Throughput 1 0,5 0 1234 Number of Cores Used Throughput (as defined by SPEC): Amount of work performed per time unit when several instances of the application is executed simultaneously. Our TP study: compare TP improvement when you go from 1 core to 4 cores PDC Summer School 2010 Institutionen för informationsteknologi | www.it.uu.se MC 18 ©Erik Hagersten| user.it.uu.se/~eh Throughput Scaling, More Apps 4,50 4,00 3,50 3,00 SPEC CPU 20006 FP Throughput improvements on 4 cores 2.02,50 2,00 1,50 1,00 0,50 0,00 PDC Summer School 410,bwaves 2010 416,gamess Intel X5365 3GHz “Core2”, 1333 MHz FSB, 8MB L2. (Based on data from the SPEC web) 433,milc Institutionen för informationsteknologi 434,zeusmp 435,gromacs 436,cactusADM 437,leslie3d 444,namd | www.it.uu.se 447,dealII 450,soplex MC 19 453,povray 454,calculix 459,GemsFDTD 465,tonto 470,lbm ©Erik Hagersten 481,wrf 482,sphinx3 | user.it.uu.se/~eh Nerd Curve: 470.LBM Miss rate (excluding HW prefetch effects) Utilization, i.e., fraction cache data
Recommended publications
  • HELLAS: a Specialized Architecture for Interactive Deformable Object Modeling
    HELLAS: A Specialized Architecture for Interactive Deformable Object Modeling Shrirang Yardi Benjamin Bishop Thomas Kelliher Department of Electrical and Department of Computing Department of Mathematics Computer Engineering Sciences and Computer Science Virginia Tech University of Scranton Goucher College Blacksburg, VA Scranton, PA Baltimore, MD [email protected] [email protected] [email protected] ABSTRACT more realistic models, it is likely that the computational Applications involving interactive modeling of deformable resources required by these simulations may continue to ex- objects require highly iterative, floating-point intensive nu- ceed those provided by the incremental advances in general- merical simulations. As the complexity of these models in- purpose systems [12]. This has created an interest in the creases, the computational power required for their simula- use of specialized hardware that provides high performance tion quickly grows beyond the capabilities of current gen- computational engines to accelerate such floating-point in- eral purpose systems. In this paper, we present the design tensive applications [6, 7]. of a low–cost, high–performance, specialized architecture to Our research focuses on the design of such specialized sys- accelerate these simulations. Our aim is to use such special- tems. However, our approach has several important differ- ized hardware to allow complex interactive physical model- ences as compared to previous work in this area: (i) we ing even on consumer-grade PCs. In this paper, we present specifically target algorithms that are numerically stable, details of the target algorithms, the HELLAS architecture, but are much more computationally intensive than those simulation results and lessons learned from our implemen- used by existing hardware implementations [3,6]; (ii) we aim tation.
    [Show full text]
  • Multiprocessing Contents
    Multiprocessing Contents 1 Multiprocessing 1 1.1 Pre-history .............................................. 1 1.2 Key topics ............................................... 1 1.2.1 Processor symmetry ...................................... 1 1.2.2 Instruction and data streams ................................. 1 1.2.3 Processor coupling ...................................... 2 1.2.4 Multiprocessor Communication Architecture ......................... 2 1.3 Flynn’s taxonomy ........................................... 2 1.3.1 SISD multiprocessing ..................................... 2 1.3.2 SIMD multiprocessing .................................... 2 1.3.3 MISD multiprocessing .................................... 3 1.3.4 MIMD multiprocessing .................................... 3 1.4 See also ................................................ 3 1.5 References ............................................... 3 2 Computer multitasking 5 2.1 Multiprogramming .......................................... 5 2.2 Cooperative multitasking ....................................... 6 2.3 Preemptive multitasking ....................................... 6 2.4 Real time ............................................... 7 2.5 Multithreading ............................................ 7 2.6 Memory protection .......................................... 7 2.7 Memory swapping .......................................... 7 2.8 Programming ............................................. 7 2.9 See also ................................................ 8 2.10 References .............................................
    [Show full text]
  • Graphics Hardware
    Department of Computer Engineering Graphics Hardware Ulf Assarsson Graphics hardware – why? l About 100x faster! l Another reason: about 100x faster! l Simple to pipeline and parallelize l Current hardware based on triangle rasterization with programmable shading (e.g., OpenGL acceleration) l Ray tracing: there are research architetures, and few commercial products – Renderdrive, RPU, (Gelato), NVIDIA OptiX 2 3 Perspective-correct texturing l How is texture coordinates interpolated over a triangle? l Linearly? Linear interpolation Perspective-correct interpolation l Perspective-correct interpolation gives foreshortening effect! l Hardware does this for you, but you need to understand this anyway! 4 5 Recall the following l Before projection, v, and after p (p=Mv) l After projection pw is not 1! l Homogenization: (px /pw , py /pw , pz /pw , 1) l Gives (px´, py ´ , pz´ , 1) 6 Texture coordinate interpolation l Linear interpolation does not work l Rational linear interpolation does: – u(x)=(ax+b) / (cx+d) (along a scanline where y=constant) – a,b,c,d are computed from triangle’s vertices (x,y,z,w,u,v) l Not really efficient l Smarter: – Compute (u/w,v/w,1/w) per vertex – These quantities can be linearly interpolated! – Then at each pixel, compute 1/(1/w)=w – And obtain: (w*u/w,w*v/w)=(u,v) – The (u,v) are perspectively-correct interpolated l Need to interpolate shading this way too – Though, not as annoying as textures l Since linear interpolation now is OK, compute, e.g., Δ(u/w)/ Δx, and use this to update u/w when stepping in the x- direction (similarly for other parameters) 7 Put differently: l Linear interpolation in screen space does not work for u,v l Solution: – We have applied a non-linear transform to each vertex (x/w, y/w, z/w).
    [Show full text]
  • Corporate Backgrounder
    AGEIA PhysX FAQs for System and Board Partners For use in partner marketing materials KEY QUESTIONS 1: What exactly is the AGEIA PhysX Processor? 2: What is the Gaming Power Triangle? 3: What is physics for gaming? 4: What is a physics processor? 5: What makes the AGEIA PhysX hardware special? 6: Why do I need physics in my game? 7: Isn’t physics just juiced up graphics? 8: Can I get PhysX enabled games today? 9: Why do I need dedicated hardware for physics? 10: How do I get the new AGEIA PhysX processor into my PC? 11: How much does AGEIA PhysX cost? 12: How do games get PhysX-optimized? 13: Do I have to buy new hardware every year just to get new physics features? AGEIA Confidential March 2006 . 1 What exactly is the AGEIA PhysX Processor? The AGEIA PhysX processor is the first dedicated hardware accelerator for PC games that enables dynamic motion and interaction for a totally new gaming experience. Exciting immersive physical gaming feature that you can expect to see in next generation cutting edge titles will include groundbreaking physics features such as: * Explosions that cause dust and collateral debris * Characters with complex, jointed geometries for more life-like motion and interaction * Spectacular new weapons with unpredictable effects * Cloth that drapes and tears the way you expect it to * Lush foliage that sways naturally when brushed against * Dense smoke & fog that billow around objects in motion What is the Gaming Power Triangle The gaming triangle represents the three critical PC processors that enable the best possible gaming experience.
    [Show full text]
  • Introduction
    A D V A N C E D C O M P U T E R G R A P H I C S Introduction (Thanks to Professions Andries van Dam and John Hughes) CMSC 635 January 15, 2013 Introduction ‹#›/16 A D V A N C E D C O M P U T E R G R A P H I C S What is Computer Graphics? • Computer graphics generally means creation, storage and manipulation of models and images • Such models come from diverse and expanding set of fields including physical, mathematical, artistic, biological, and even conceptual (abstract) structures Frame from animation by William Latham, shown at SIGGRAPH 1992. Latham uses rules that govern patterns of natural forms to create his artwork. CMSC 635 January 15, 2013 Introduction ‹#›/16 A D V A N C E D C O M P U T E R G R A P H I C S Modeling vs. Rendering • Modeling – Create models – Apply materials to models – Place models around scene – Place lights in scene – Place the camera • Rendering – Take “picture” with camera • Both can be done by modern commercial software: – Autodesk MayaTM ,3D Studio MaxTM, BlenderTM, etc. Point Light Spot Light Directional Light Ambient Light CS128 lighting assignment by Patrick Doran, Spring 2009 CMSC 635 January 15, 2013 Introduction ‹#›/16 A D V A N C E D C O M P U T E R G R A P H I C S Solved problem: Rendering Avatar 2009 CMSC 635 January 15, 2013 Introduction ‹#›/16 A D V A N C E D C O M P U T E R G R A P H I C S What is Interactive Computer Graphics? (1/2) • User controls contents, structure, and appearance of objects and their displayed images via rapid visual feedback • Basic components of an interactive graphics system – input (e.g., mouse, tablet and stylus, multi-touch…) – processing (and storage) – display/output (e.g., screen, paper-based printer, video recorder…) • First truly interactive graphics system, Sketchpad, pioneered at MIT by Ivan Sutherland for his 1963 Ph.D.
    [Show full text]
  • Gputerasort: High Performance Graphics Co-Processor Sorting For
    GPUTeraSort: High Performance Graphics Co­processor Sorting for Large Database Management Naga K. Govindaraju ∗ Jim Gray † Ritesh Kumar ∗ Dinesh Manocha ∗ naga,ritesh,dm @cs.unc.edu, [email protected] { } http://gamma.cs.unc.edu/GPUTERASORT ABSTRACT advanced memory units, and multi-processors to improve sorting performance. The current Indy PennySort record We present a new algorithm, GPUTeraSort, to sort billion- 1 record wide-key databases using a graphics processing unit benchmark , sorts a 40 GB database in 1541 seconds on a (GPU) Our algorithm uses the data and task parallelism $614 Linux/AMD system. on the GPU to perform memory-intensive and compute- However, current external memory sort performance is intensive tasks while the CPU is used to perform I/O and limited by the traditional Von Neumann style architecture resource management. We therefore exploit both the high- of the CPU. Computer architects use data caches to amelio- bandwidth GPU memory interface and the lower-bandwidth rate the CPU and the main memory bottleneck; but, CPU- CPU main memory interface and achieve higher memory based sorting algorithms incur significant cache misses on bandwidth than purely CPU-based algorithms. GPUTera- large datasets. Sort is a two-phase task pipeline: (1) read disk, build keys, This article shows how to use a commodity graphics pro- sort using the GPU, generate runs, write disk, and (2) read, cessing unit (GPU) as a co-processor to sort large datasets. merge, write. It also pipelines disk transfers and achieves GPUs are programmable parallel architectures designed for near-peak I/O performance. We have tested the perfor- real-time rasterization of geometric primitives - but they mance of GPUTeraSort on billion-record files using the stan- are also highly parallel vector co-processors.
    [Show full text]
  • The Hardware Accelerated Physics Engine with Operating Parameters Controlling the Numerical Error Tolerance 1
    International Journal of Advanced Science and Technology Vol.119 (2018), pp.145-152 http://dx.doi.org/10.14257/ijast.2018.119.13 The Hardware Accelerated Physics Engine with Operating Parameters Controlling the Numerical Error Tolerance 1 Youngsik Kim Department of Game and Multimedia Engineering, Korea Polytechnic University [email protected] Abstract This paper designed the dedicated hardware pipeline for interactive 3D entertainment system considering parallelization characteristics of rigid body physics simulation. This paper also analyzed the large parallelization characteristics of independent rigid group units connected to joints in hierarchical collision detection and design a hardware structure suitable for each parallel processing stage by extracting small data parallelization characteristics of each constraint row unit of LCP equation. Also, this paper performed control such as the number of loop repetitions by performing hardware trade-off in which numerical error is reduced within visual error tolerance by setting error attenuation parameter (ERP) and constraint force mixing (CFM) as driving parameters. Keywords: Hardware Accelerator, Physics Engine, Open Dynamics Engine, Numerical Error Tolerance 1. Introduction The importance of physics engine for simulating the motion of an object about physical phenomena (mass, speed, friction, collision) of nature in the interactive 3D game, virtual reality, 3D animation and robotics industry is becoming more and more important [1-4]. In a car racing game, it is better to use a physics engine simulation than a keyframe animation created by an artist, as it slips by centrifugal force when cornering and collides head-on with other cars. It would be much more natural to model an automobile as a rigid body with mass, a moment of inertia, velocity, etc., and simulate the kinematic motion of the body due to collision or friction.
    [Show full text]
  • F06 Physics Processor
    Nicholas Bannister, Vinay Chaudhary, Aaron Hoy, Charles Norman, Jeremy Weagley Team VerchuCom PPU + Demo Final Report What We Built Game Description Overall Implementation Our original plan of action was to implement the game of Worms that utilized a customized Physics Processing Unit (PPU) built on the FPGA. We had planned to run the game on the Workstation, send objects back and forth to a shared bank of memory on the board, and update the objects’ state using the PPU. However, we came to realize that Worms would not fully show off our main feature of the project, which is the PPU. Therefore, we decided to create a “demo” which basically consisted of a bunch of squares interacting with each other, wind resistance and gravity within a confined area. We then took it to the next level and implemented a simple 2-player shooter game. The object of the game is for each player to destroy all of their squares, either blue or square, before their opponent. Each person has unlimited regular ammo along with 2 “nuclear bombs,” which destroy all of the bombs within a certain radius. Physics Processing Unit (PPU) We designed a specialized CPU (A PPU) that operates on structs representing 2 dimensional objects (position, velocity, acceleration, orientation, etc.). It modeled various particle effects, such as • Model collisions between particles and elastic/inelastic collisions • Effects of Gravity • Wind Resistance The accelerator sits in a basic iteration loop that applies each of the various effects on all the particles in memory, updating their positions, velocities, etc. based on physics.
    [Show full text]
  • PC Hardware Contents
    PC Hardware Contents 1 Computer hardware 1 1.1 Von Neumann architecture ...................................... 1 1.2 Sales .................................................. 1 1.3 Different systems ........................................... 2 1.3.1 Personal computer ...................................... 2 1.3.2 Mainframe computer ..................................... 3 1.3.3 Departmental computing ................................... 4 1.3.4 Supercomputer ........................................ 4 1.4 See also ................................................ 4 1.5 References ............................................... 4 1.6 External links ............................................. 4 2 Central processing unit 5 2.1 History ................................................. 5 2.1.1 Transistor and integrated circuit CPUs ............................ 6 2.1.2 Microprocessors ....................................... 7 2.2 Operation ............................................... 8 2.2.1 Fetch ............................................. 8 2.2.2 Decode ............................................ 8 2.2.3 Execute ............................................ 9 2.3 Design and implementation ...................................... 9 2.3.1 Control unit .......................................... 9 2.3.2 Arithmetic logic unit ..................................... 9 2.3.3 Integer range ......................................... 10 2.3.4 Clock rate ........................................... 10 2.3.5 Parallelism .........................................
    [Show full text]
  • A Comprehensive Survey of Various Processor Types & Latest
    International Journal of Research and Scientific Innovation (IJRSI) | Volume V, Issue IV, April 2018 | ISSN 2321–2705 A Comprehensive Survey of Various Processor types & Latest Architectures Harisha M.S1, Dr. D. Jayadevappa2 1Research Scholar, Jain University, Bangalore, Karnataka, India 2Professor, Electronics Instrumentation, Dept., JSSATE, Bangalore, Karnataka, India Abstract- This technology survey paper covers application based Processors- ASIP, Processor based on Flynn’s classification which includes SISD, SIMD, MISD, MIMD, Special processors like Graphics processing unit (GPU), Physics processing unit (PPU), Digital signal processor (DSP), network processor, front end processor, co-processor and processor based on number of cores which includes- single core, multi core, multi- processor, hyper threading and multi core with shared cache processors. Keywords- ASIP, SISD, SIMD, MISD, MIMD, GPU, PPU, DSP, Network processor, coprocessor. I. INTRODUCTION A. Types of processors- Processor based on application Application-specific instruction-set processor (ASIP) Application-specific instruction set processor is a component Fig.2 Flynn’s classification used in system-on-a-chip design. The instruction set of an SISD & SIMD ASIP is designed to benefit a specific application [1]. This specialization of the core provides a tradeoff between the flexibility of a general purpose CPU and the performance of an ASIC. Fig.3 SISD and SIMD SISD SISD is a computer architecture in which a single uni-core processor executes a single instruction . Fig.1 ASIP stream, to operate on data stored in a single memory. A sequential computer which exploits no parallelism B. Processor types based on Flynn’s classification in either the instruction or data streams. In Flynn's taxonomy, the processors are classified based on Examples of SISD architecture are the traditional the number of concurrent instruction and data streams uniprocessor machines like PC or old mainframes available in the architecture: Single Instruction Single Data [3].
    [Show full text]
  • Revisiting the Future ☺
    Revisiting the Future ☺ Erik Hagersten Uppsala University Sweden DARK2 in a nutshell 1. Memory Systems (caches, VM, DRAM, microbenchmarks, …) 2. Multiprocessors (TLP, coherence, interconnects, scalability, clusters, …) 3. CPUs (pipelines, ILP, scheduling, Superscalars, VLIWs, embedded, …) 4. Future: (physical limitations, TLP+ILP DARK in the CPU,…) 2009 Dept of Information Technology| www.it.uu.se 2 ©Erik Hagersten| user.it.uu.se/~eh How do we get good performance? Creating and exploring: 1) Locality a) Spatial locality b) Temporal locality c) Geographical locality 2) Parallelism a) Instruction level (ILP) b) Thread level (TLP) DARK c) Memory level (MLP) 2009 Dept of Information Technology| www.it.uu.se 3 ©Erik Hagersten| user.it.uu.se/~eh Ray Kurzweil pictures www.KurzweilAI.net/pps/WorldHealthCongress/ DARK 2009 Dept of Information Technology| www.it.uu.se 4 ©Erik Hagersten| user.it.uu.se/~eh Ray Kurzweil pictures www.KurzweilAI.net/pps/WorldHealthCongress/ DARK 2009 Dept of Information Technology| www.it.uu.se 5 ©Erik Hagersten| user.it.uu.se/~eh Ray Kurzweil pictures www.KurzweilAI.net/pps/WorldHealthCongress/ DARK 2009 Dept of Information Technology| www.it.uu.se 6 ©Erik Hagersten| user.it.uu.se/~eh Ray Kurzweil pictures www.KurzweilAI.net/pps/WorldHealthCongress/ DARK 2009 Dept of Information Technology| www.it.uu.se 7 ©Erik Hagersten| user.it.uu.se/~eh Doubling (or Halving) times Dynamic RAM Memory (bits per dollar) 1.5 years Average Transistor Price 1.6 years Microprocessor Cost per Transistor Cycle 1.1 years Total Bits Shipped
    [Show full text]
  • System-On-A-Chip
    System-on-a-chip From Wikipedia, the free encyclopedia Jump to: navigation, search System-on-a-chip or system on chip (SoC or SOC) is an idea of integrating all components of a computer or other electronic system into a single integrated circuit (chip). It may contain digital, analog, mixed-signal, and often radio-frequency functions – all on one chip. A typical application is in the area of embedded systems. If it is not feasible to construct an SoC for a particular application, an alternative is a system in package (SiP) comprising a number of chips in a single package. SoC is believed to be more cost effective since it increases the yield of the fabrication and because its packaging is simpler. Contents [hide] • 1 Structure • 2 Design flow • 3 Fabrication • 4 See also • 5 External links [edit] Structure y513719001187192499 from [email protected] was published by D-Publish on August 15, 2007 Microcontroller-based System-on-a-Chip A typical SoC consists of: • One or more microcontroller, microprocessor or DSP core(s). • Memory blocks including a selection of ROM, RAM, EEPROM and Flash. • Timing sources including oscillators and phase-locked loops. • Peripherals including counter-timers, real-time timers and power-on reset generators. • External interfaces including industry standards such as USB, FireWire, Ethernet, USART, SPI. • Analog interfaces including ADCs and DACs. • Voltage regulators and power management circuits. These blocks are connected by either a proprietary or industry-standard bus such as the AMBA bus from ARM. DMA controllers route data directly between external interfaces and memory, by-passing the processor core and thereby increasing the data throughput of the SoC.
    [Show full text]