Semiconductor Memories: I T D Ti an Introduction

Total Page:16

File Type:pdf, Size:1020Kb

Semiconductor Memories: I T D Ti an Introduction Semiconductor Memories: an ItIntrod ucti on Talk Overview MTdMemory Trend Memory Classification Memory Architectures The Memory Core PihPeriphery Reliability Semiconductor Memory Trends (up to the 90’s ) Memory Size as a function of time: x 4 every three years Semiconductor Memory Trends (Upd at ed Furth er Beyond) Trends in Memory Cell Area Growth in DRAM Chip Capacity 1000000 256, 000 100000 64,000 yy 16,000 10000 pacit 4,000 aa 1000 1,000 Kbit c Kbit 256 100 64 10 1980 1982 1984 1986 1988 1990 1992 1994 1996 1998 2000 Year of introduction Semiconductor Memory Classification Non-Volatile Read-Write Memory Read-Write Read-Only Memory Memory Random Non-Random EPROM Mask-Programmed Access Access 2 E PROM Programmable (PROM) DRAM FIFO FLASH SRAM LIFO MRAM Shift Register PRAM RRAM CAM Memory Timing: Definitions Read cycle READ Write cycle Read access Read access WRITE Write access Data valid DATA Data written Memory Architecture: Decoders Intuitive architecture for n x m memory Too many select signals: N words == N select signals m bits m bits S S0 Word 0 0 Word 0 S S1 Word 1 1 Word 1 S2 S2 Word 2 Storage A0 Word 2 Storage Cell S3 Cell S3 A1 Ak-1 S Sn-2 n-2 Word n-2 Word n-2 S Sn-1 n-1 Word n-1 Word n-1 Input/Output Input/Output n words → n select signals Decoder reduces # of inputs k = log2 n ArrayArray--StructuredStructured Memory Architecture 2k-j bit line word line Aj A j+1 storage (RAM) cell Ak-1 m2j A 0 selects appropriate word A1 Column Decoder from memory row Aj-1 Sense Amplifiers amplifies bit line swing Read/Write Circuits Input/Output (m bits) Hierarchical Memory Architecture Block 0 Block i Block P 21- Row address ClColumn address Block address Global data bus Control Block selector Global circuitry amplifier/driver I/O Advantages: 1. Shorter wires within blocks for reduced local transit times 2. Block address activates only 1 block for power savings Block Diagram of 4 Mbit SRAM Clock Z-address X-address generator buffer buffer Predecoder and block selector Bit line load cccc cccc row dec row dec al row al row dede al row de al row de llll b bbb glo glob -glo -glob GlobaGloba SubSub- SubSub- Transfer gate Column decoder Sense amplifier and write driver Local row dec CS, WE I/O x1/x4 Y-address X-address buffer buffer controller buffer buffer Memory Timing: Approaches Address bus Row Address Column Address RAS Address Bus Address Address transition CAS initiates memory operation RAS-CAS timing DRAM Timing: SRAM Timing: MltilMultiplexed dAd Adressi ng SlfSelf-time d Read-Only Memory Cells BL BL BL VDD WL WL WL 1 BL BL BL WL WL WL 0 GND Diode ROM MOS ROM 1 MOS ROM 2 MOS OR ROM BL[0] BL[1] BL[2] BL[3] WL[0] VDD WL[1] WL[2] VDD WL[3] Vbias Pull-down loads MOS NOR ROM VDD Pull-up devices WL[0] GND WL [1] WL [2] GND WL [3] BL [[]0] BL [[]1] BL [[]2] BL [[]3] MOS NAND ROM VDD Pull-up devices BL[0] BL[1] BL[2] BL[3] WL[0] WL[1] WL[2] WL[3] All word li nes hi gh b y d ef ault with excepti on of sel ect ed row Equiva le nt Tra ns ie n t M odel f or M OS N OR R OM VDD BL r WL word Cbit cword Word line parasitics Wire capacitance and gate capacitance Wire resistance (polysilicon) Bit line parasiiitics Resistance not dominant (metal) Drain jjgunction and gate-drain overlappp capacitance Equivalent Transient Model for MOS NAND ROM VDD BL CL rbit c r bit WL word cword Word line parasitics Similar to NOR ROM Bit line parasitics Resistance of cascaded transistors dominates. Drain/source and complete gate capacitances Decreasing Word Line Delay Drive the word line from both sides driver driver polysilicon word line WL metal word line Use a metal bypass polysilicon word line WL metal bypass Use silicides Precharged MOS NOR ROM V f pre DD Precharge devices WL[0] GND WL[1] WL[2] GND WL[3] BL[0] BL[[]1] BL[[]2] BL[3] PMOS precharge device can be made as large as necessary, but clock driver becomes harder to design. NonNon--VolatileVolatile Memories The Float ing-gate transistor (FAMOS) Floating gate Gate D Source Drain tox G tox S n+ p n+_ Substrate Device cross-section Schematic symbol Floating-Gate Transistor Programming 20 V 0 V 5 V 10 V 5 V 20 V 5 V 0 V 2.5 V 5 V S D S D S D Avalanche injection Removing programming Programming results in voltage leaves ch arge tdtrapped. hig her VT. A “Programmable-Threshold” Transistor I D “0”-state “1”-state “ ON” ΔDVT “ OFF” VWL VGS Floating-Gate Tunneling Oxide (FLOTOX) EEPROM Floating gate Gate I Source Drain 20–30 nm -10 V VGD 10 V + + nn1 p n1 Substrate 10 nm Fowler-Nordheim FLOTOX transistor I-V characteristic EEPROM Cell BL WL Absolute threshold control is hard, and non- ppgrogrammed transistor might be in depletion. VDD Ö 2-transistor cell (one serving as the access transistor) Flash EEPROM Control gate Floating gate erasure Thin tunneling oxide +1 n+1 nn source programming n drain p-substrate Erasure using Fowler-Nordheim tunneling is performed in bulk for the complete chip or in a sub-section of the memory. Cross-sections of NVM cells Flhlash EPROM Courtesy Intel Basic Operations in a NOR Flash Memory― Erase cell array BL 0 BL 1 G 12 V 0 V WL 0 S D 12 V 0 V WL 1 open open Basic Operations in a NOR Flash Memory― Write 12 V BL 0 BL 1 G 6 V 12 V WL 0 S D 0 V 0V0 V WL 1 6 V 0 V Basic Operations in a NOR Flash Memory― Read BL BL 5 V 0 1 G 1V1 V 5 V WL 0 S D 0 V 0V0 V WL 1 1V1 V 0V0 V NAND Flash Memory Word line(poly) Unit Cell Gate ONO Gate FG Oxide Source line (Diff. Layer) Courtesy Toshiba NAND Flash Memory Select transistor Word lines Active area STI Bit line contact Source line contact Read-Write Memories (RAM) STATIC (SRAM) Data stored as long as power supply is applied Large (6 transistors/cell) Fast Differential DYNAMIC (DRAM) Periodic refresh required Small (1-3 transistors/cell) Slower Single Ended 6-transistor CMOS SRAM Cell WL VDD M 2 M 4 Q Q M M 5 6 M 1 M 3 BL BL CMOS SRAM Analysis (Read) WL VDD M BL 4 BL Q = 0 M Q 1 6 M 5 = V VDD M1 DD VDD Cbit Cbit CMOS SRAM Analysis (Read) 121.2 1 V) (( V 0.8 Δ 0.6 se, ii 0.4 ltage R 0.2 oo V 0 0 0.5 1 1.2 1.5 2 2.5 3 Cell Ratio (CR) CMOS SRAM Analysis (Write) WL VDD M4 Q = 0 M6 M 5 Q = 1 M1 VDD BL = 1 BL = 0 CMOS SRAM Analysis (Write) W / L PR 4 4 W6 / L6 Resistance-load SRAM Cell WL VDD RL RL QQ M3 M 4 BL BL M 1 M 2 Static power dissipation -- Want R L large Bit lines precharged to VDD to address tp problem 3-Transistor DRAM Cell BL1 BL2 WWL RWL WWL M 3 RWL M 1 X X V DD 2-V T M 2 V DD CS BL 1 ΔD V BL 2 V DD 2- V T No constraints on device ratios Reads are non-destructive. Value stored at node X when writing a “1” = VWWL-VTn 1-Transistor DRAM Cell BL WL Write 1 Read 1 WL M 1 X X GND VDD 2- VT CS VDD BL V /2 V /2 DD sensing DD CBL Write: CS is charged or discharged by asserting WL and BL. Read: Charge redistribution takes places between bit line and storage capacitance Voltage swing is small, typically around 250 mV. C ------------S ΔV ==VBL – VPRE (V X – V PRE ) CS + CBL DRAM Cell Observations 1T DRAM requires a sense amplifier for each bit line due thto charge redi ditibtistribution read-out. The read-out of the 1T DRAM cell is destructive; read and refresh operations are necessary for correct operation. DRAM memory cells are single-ended in contrast to SRAM cells. Unlike 3T cell, 1T cell requires presence of an extra capacitance that must be explicitly included in the design. When writingg,g a “1” into a DRAM cell, a threshold voltage is lost. This charge loss can be circumvented by bootstrapping the word lines to a higher value than VDD. Sense Amplifier Operation VBL V(1) VPRE DΔV(1) V(0) Sense amp activated t Word line activated 1-T DRAM Cell Capacitor M1 word line Metal word line SiO2 Poly n+ n+ Field Oxide Diffuse d bit line Inversion layer Poly Polysilicon induced by Polysilicon plate bias gate plate Cross-section Layout Uses Polysilicon-Diffusion Capacitance Expensive in Area Advanced 1T DRAM Cells Word line Capacitor dielectric layer Insulating Layer Cell plate Cell Plate Si Transfer gate Isolation Refillinggy Poly Capacitor Insulator Storage electrode Storage Node Poly Si Substrate 2nd Field Oxide ThCllTrench Cell Stacked Cell Row Decoders Collection of 2M complex logggic gates Organized in regular and dense fashion (N)AND Decoder NOR Decoder Hierarchical Decoders Multi-stage implementation improves performance ••• WL 1 WL 0 A 0A 1 A 0A 1 A 0A 1 A 0A 1 A 2A 3 A 2A 3 A 2A 3 A 2A 3 ••• NAND decoder using 22--inputinput prepre--decodersdecoders A 1 A 0 A 0 A 1 A 3 A 2 A 2 A 3 Dynamic Decoders PhPrecharge didevices GND GND VDD WL3 VDD WL3 WL WL 2 2 VDD WL1 WL 1 VDD WL0 WL 0 VDD φ A0 A0 A1 A1 A0 A0 A1 A1 φ 2-input NOR decoder 2-input NAND decoder 4-to-1 tree based column decoder BL 0 BL 1 BL 2 BL 3 A 0 A 0 A1 A 1 D Number of devices drastically reduced Delay increases quadratically with # of sections; prohibitive for large decoders Solutions: buffers progressive sizing combinati on o f t ree an d pass t ransi st or approach es Sense Amplifiers make ΔVassmallV as small C×ΔV as possible (make the SA tp = ---------------- Iav as sensitive as possible) large small Idea: Use Sense Amplifer small transition s.a.
Recommended publications
  • File Systems and Disk Layout I/O: the Big Picture
    File Systems and Disk Layout I/O: The Big Picture Processor interrupts Cache Memory Bus I/O Bridge Main I/O Bus Memory Disk Graphics Network Controller Controller Interface Disk Disk Graphics Network 1 Rotational Media Track Sector Arm Cylinder Platter Head Access time = seek time + rotational delay + transfer time seek time = 5-15 milliseconds to move the disk arm and settle on a cylinder rotational delay = 8 milliseconds for full rotation at 7200 RPM: average delay = 4 ms transfer time = 1 millisecond for an 8KB block at 8 MB/s Bandwidth utilization is less than 50% for any noncontiguous access at a block grain. Disks and Drivers Disk hardware and driver software provide basic facilities for nonvolatile secondary storage (block devices). 1. OS views the block devices as a collection of volumes. A logical volume may be a partition ofasinglediskora concatenation of multiple physical disks (e.g., RAID). 2. OS accesses each volume as an array of fixed-size sectors. Identify sector (or block) by unique (volumeID, sector ID). Read/write operations DMA data to/from physical memory. 3. Device interrupts OS on I/O completion. ISR wakes up process, updates internal records, etc. 2 Using Disk Storage Typical operating systems use disks in three different ways: 1. System calls allow user programs to access a “raw” disk. Unix: special device file identifies volume directly. Any process that can open thedevicefilecanreadorwriteany specific sector in the disk volume. 2. OS uses disk as backing storage for virtual memory. OS manages volume transparently as an “overflow area” for VM contents that do not “fit” in physical memory.
    [Show full text]
  • Lecture 27 Semiconductor Memory: DRAM and Non-Volatile Memory Administrivia
    Lecture 27 Semiconductor Memory: DRAM and Non-Volatile Memory Digital Integrated Circuits Interconnect © Prentice Hall 2000 Administrivia l Today: Project phase 3 announcement. l Poster Session Tu 5/8 1-4pm » Location BWRC, 2108 Allston Way l Last lecture on Th 5/3 will cover issues in IC design Digital Integrated Circuits Interconnect © Prentice Hall 2000 1 Lectures Last l ROM and SRAM Today l Introducing the project phase III l DRAM and Non-volatile Digital Integrated Circuits Interconnect © Prentice Hall 2000 Project Phase III A proposed SRAM cell! w/ Control Circuit Digital Integrated Circuits Interconnect © Prentice Hall 2000 2 Tasks l Explain the behavior of the cell in its global contents. Provide transient simulations to illustrate. l Identify weakness of the cell in terms of signal integrity and power dissipation. Quantify your statements. l Propose an implementation that improves power dissipation. Digital Integrated Circuits Interconnect © Prentice Hall 2000 Report l Report » Comparison, Selection, Electrical Design » Cell Layout, Timing Waveforms in SRAM, Simulation » Power and Estimation and Proposal for SRAM power reduction l 3 slides on poster, each of which represents one of the tasks of the previous slide » Explanation of cell operation, comparison, design » Cell Operation in SRAM, Waveforms, Simulation » Proposal of improved SRAM implementation (from a power perspective) Digital Integrated Circuits Interconnect © Prentice Hall 2000 3 SEMICONDUCTOR MEMORIES Digital Integrated Circuits Interconnect © Prentice Hall 2000 3-Transistor DRAM Cell BL1 BL2 WWL WWL RWL RWL X X M3 VDD-VT M2 M1 VDD BL1 CS BL2 VDD-VT DV No constraints on device ratios Reads are non-destructive Value stored at node X when writing a “1” = VWWL-VTn Digital Integrated Circuits Interconnect © Prentice Hall 2000 4 3T-DRAM — Layout BL2 BL1 GND RWL M3 M2 WWL M1 Digital Integrated Circuits Interconnect © Prentice Hall 2000 1-Transistor DRAM Cell BL WL Write "1" Read "1" WL 1 M C X S GND VDD-VT V BL DD VDD/2 VDD/2 CBL sensing Write: CS is charged or discharged by asserting WL and BL.
    [Show full text]
  • Nanotechnology Trends in Nonvolatile Memory Devices
    IBM Research Nanotechnology Trends in Nonvolatile Memory Devices Gian-Luca Bona [email protected] IBM Research, Almaden Research Center © 2008 IBM Corporation IBM Research The Elusive Universal Memory © 2008 IBM Corporation IBM Research Incumbent Semiconductor Memories SRAM Cost NOR FLASH DRAM NAND FLASH Attributes for universal memories: –Highest performance –Lowest active and standby power –Unlimited Read/Write endurance –Non-Volatility –Compatible to existing technologies –Continuously scalable –Lowest cost per bit Performance © 2008 IBM Corporation IBM Research Incumbent Semiconductor Memories SRAM Cost NOR FLASH DRAM NAND FLASH m+1 SLm SLm-1 WLn-1 WLn WLn+1 A new class of universal storage device : – a fast solid-state, nonvolatile RAM – enables compact, robust storage systems with solid state reliability and significantly improved cost- performance Performance © 2008 IBM Corporation IBM Research Non-volatile, universal semiconductor memory SL m+1 SL m SL m-1 WL n-1 WL n WL n+1 Everyone is looking for a dense (cheap) crosspoint memory. It is relatively easy to identify materials that show bistable hysteretic behavior (easily distinguishable, stable on/off states). IBM © 2006 IBM Corporation IBM Research The Memory Landscape © 2008 IBM Corporation IBM Research IBM Research Histogram of Memory Papers Papers presented at Symposium on VLSI Technology and IEDM; Ref.: G. Burr et al., IBM Journal of R&D, Vol.52, No.4/5, July 2008 © 2008 IBM Corporation IBM Research IBM Research Emerging Memory Technologies Memory technology remains an
    [Show full text]
  • Semiconductor Memories
    Semiconductor Memories Prof. MacDonald Types of Memories! l" Volatile Memories –" require power supply to retain information –" dynamic memories l" use charge to store information and require refreshing –" static memories l" use feedback (latch) to store information – no refresh required l" Non-Volatile Memories –" ROM (Mask) –" EEPROM –" FLASH – NAND or NOR –" MRAM Memory Hierarchy! 100pS RF 100’s of bytes L1 1nS SRAM 10’s of Kbytes 10nS L2 100’s of Kbytes SRAM L3 100’s of 100nS DRAM Mbytes 1us Disks / Flash Gbytes Memory Hierarchy! l" Large memories are slow l" Fast memories are small l" Memory hierarchy gives us illusion of large memory space with speed of small memory. –" temporal locality –" spatial locality Register Files ! l" Fastest and most robust memory array l" Largest bit cell size l" Basically an array of large latches l" No sense amps – bits provide full rail data out l" Often multi-ported (i.e. 8 read ports, 2 write ports) l" Often used with ALUs in the CPU as source/destination l" Typically less than 10,000 bits –" 32 32-bit fixed point registers –" 32 60-bit floating point registers SRAM! l" Same process as logic so often combined on one die l" Smaller bit cell than register file – more dense but slower l" Uses sense amp to detect small bit cell output l" Fastest for reads and writes after register file l" Large per bit area costs –" six transistors (single port), eight transistors (dual port) l" L1 and L2 Cache on CPU is always SRAM l" On-chip Buffers – (Ethernet buffer, LCD buffer) l" Typical sizes 16k by 32 Static Memory
    [Show full text]
  • EEPROM Emulation
    ...the world's most energy friendly microcontrollers EEPROM Emulation AN0019 - Application Note Introduction This application note demonstrates a way to use the flash memory of the EFM32 to emulate single variable rewritable EEPROM memory through software. The example API provided enables reading and writing of single variables to non-volatile flash memory. The erase-rewrite algorithm distributes page erases and thereby doing wear leveling. This application note includes: • This PDF document • Source files (zip) • Example C-code • Multiple IDE projects 2013-09-16 - an0019_Rev1.09 1 www.silabs.com ...the world's most energy friendly microcontrollers 1 General Theory 1.1 EEPROM and Flash Based Memory EEPROM stands for Electrically Erasable Programmable Read-Only Memory and is a type of non- volatile memory that is byte erasable and therefore often used to store small amounts of data that must be saved when power is removed. The EFM32 microcontrollers do not include an embedded EEPROM module for byte erasable non-volatile storage, but all EFM32s do provide flash memory for non-volatile data storage. The main difference between flash memory and EEPROM is the erasable unit size. Flash memory is block-erasable which means that bytes cannot be erased individually, instead a block consisting of several bytes need to be erased at the same time. Through software however, it is possible to emulate individually erasable rewritable byte memory using block-erasable flash memory. To provide EEPROM functionality for the EFM32s in an application, there are at least two options available. The first one is to include an external EEPROM module when designing the hardware layout of the application.
    [Show full text]
  • A Hybrid Swapping Scheme Based on Per-Process Reclaim for Performance Improvement of Android Smartphones (August 2018)
    Received August 19, 2018, accepted September 14, 2018, date of publication October 1, 2018, date of current version October 25, 2018. Digital Object Identifier 10.1109/ACCESS.2018.2872794 A Hybrid Swapping Scheme Based On Per-Process Reclaim for Performance Improvement of Android Smartphones (August 2018) JUNYEONG HAN 1, SUNGEUN KIM1, SUNGYOUNG LEE1, JAEHWAN LEE2, AND SUNG JO KIM2 1LG Electronics, Seoul 07336, South Korea 2School of Software, Chung-Ang University, Seoul 06974, South Korea Corresponding author: Sung Jo Kim ([email protected]) This work was supported in part by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education under Grant 2016R1D1A1B03931004 and in part by the Chung-Ang University Research Scholarship Grants in 2015. ABSTRACT As a way to increase the actual main memory capacity of Android smartphones, most of them make use of zRAM swapping, but it has limitation in increasing its capacity since it utilizes main memory. Unfortunately, they cannot use secondary storage as a swap space due to the long response time and wear-out problem. In this paper, we propose a hybrid swapping scheme based on per-process reclaim that supports both secondary-storage swapping and zRAM swapping. It attempts to swap out all the pages in the working set of a process to a zRAM swap space rather than killing the process selected by a low-memory killer, and to swap out the least recently used pages into a secondary storage swap space. The main reason being is that frequently swap- in/out pages use the zRAM swap space while less frequently swap-in/out pages use the secondary storage swap space, in order to reduce the page operation cost.
    [Show full text]
  • AN-1471 Application Note
    AN-1471 APPLICATION NOTE One Technology Way • P. O. Box 9106 • Norwood, MA 02062-9106, U.S.A. • Tel: 781.329.4700 • Fax: 781.461.3113 • www.analog.com ADuCM4050 Flash EEPROM Emulation by Pranit Jadhav and Rafael Lajara INTRODUCTION provides assurance that the flash initialization function works as Nonvolatile data storage is a necessity in many embedded expected. An ECC check is enabled for the entire user space in systems. Data, such as boot up configuration, calibration constants, the flash memory of the ADuCM4050. If ECC errors are reported and network related information, is normally stored on an during any read operation, the ECC engine automatically corrects electronically erasable programmable read only memory 1-bit errors and only reports on detection of 2-bit errors. When a (EEPROM) device. The advantage of using the EEPROM to store read occurs on the flash, appropriate flags are set in the status this data is that a single byte on an EEPROM device can be register of the ADuCM4050. If interrupts are generated, the source rewritten or updated without affecting the contents in the other address of the ECC error causing an interrupt is available in the locations. FLCC0_ECC_ADDR register for the interrupt service routine (ISR) to read. The ADuCM4050 is an ultra low power microcontroller unit (MCU) with integrated flash memory. The ADuCM4050 includes Emulation of the EEPROM on the integrated flash memory 512 kB of embedded flash memory with a 72-bit wide data bus reduces the bill of material (BOM) cost by omitting the EEPROM that provides two 32-bits words of data and one corresponding in the design.
    [Show full text]
  • File System Layout
    File Systems Main Points • File layout • Directory layout • Reliability/durability Named Data in a File System index !le name directory !le number structure storage o"set o"set block Last Time: File System Design Constraints • For small files: – Small blocks for storage efficiency – Files used together should be stored together • For large files: – ConCguous allocaon for sequenCal access – Efficient lookup for random access • May not know at file creaon – Whether file will become small or large File System Design Opons FAT FFS NTFS Index Linked list Tree Tree structure (fixed, asym) (dynamic) granularity block block extent free space FAT array Bitmap Bitmap allocaon (fixed (file) locaon) Locality defragmentaon Block groups Extents + reserve Best fit space defrag MicrosoS File Allocaon Table (FAT) • Linked list index structure – Simple, easy to implement – SCll widely used (e.g., thumb drives) • File table: – Linear map of all blocks on disk – Each file a linked list of blocks FAT MFT Data Blocks 0 1 2 3 !le 9 block 3 4 5 6 7 8 9 !le 9 block 0 10 !le 9 block 1 11 !le 9 block 2 12 !le 12 block 0 13 14 15 16 !le 12 block 1 17 18 !le 9 block 4 19 20 FAT • Pros: – Easy to find free block – Easy to append to a file – Easy to delete a file • Cons: – Small file access is slow – Random access is very slow – Fragmentaon • File blocks for a given file may be scaered • Files in the same directory may be scaered • Problem becomes worse as disk fills Berkeley UNIX FFS (Fast File System) • inode table – Analogous to FAT table • inode – Metadata • File owner, access permissions,
    [Show full text]
  • ZFS On-Disk Specification Draft
    ZFS On-Disk Specification Draft Sun Microsystems, Inc. 4150 Network Circle Santa Clara, CA 95054 U.S.A 1 ©2006 Sun Microsystems, Inc. 4150 Network Circle Santa Clara, CA 95054 U.S.A. This product or document is protected by copyright and distributed under licenses restricting its use, copying, distribution, and decompilation. No part of this product or document may be reproduced in any form by any means without prior written authorization of Sun and its licensors, if any. Third-party software, including font technology, is copyrighted and licensed from Sun suppliers. Parts of the product may be derived from Berkeley BSD systems, licensed from the University of California. Sun, Sun Microsystems, the Sun logo, Java, JavaServer Pages, Solaris, and StorEdge are trademarks or registered trademarks of Sun Microsystems, Inc. in the U.S. and other countries. U.S. Government Rights Commercial software. Government users are subject to the Sun Microsystems, Inc. standard license agreement and applicable provisions of the FAR and its supplements. DOCUMENTATION IS PROVIDED AS IS AND ALL EXPRESS OR IMPLIED CONDITIONS, REPRESENTATIONS AND WARRANTIES, INCLUDING ANY IMPLIED WARRANTY OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT, ARE DISCLAIMED, EXCEPT TO THE EXTENT THAT SUCH DISCLAIMERS ARE HELD TO BE LEGALLY INVALID. Unless otherwise licensed, use of this software is authorized pursuant to the terms of the license found at: http://developers.sun.com/berkeley_license.html Ce produit ou document est protégé par un copyright et distribué avec des licences qui en restreignent l'utilisation, la copie, la distribution, et la décompilation. Aucune partie de ce produit ou document ne peut être reproduite sous aucune forme, par quelque moyen que ce soit, sans l'autorisation préalable et écrite de Sun et de ses bailleurs de licence, s'il y en a.
    [Show full text]
  • 1. Direct Mapped Caches 1
    CS 61C SpringFall 2014 2014 – Discussion 5 – Direct Mapped Caches In the following diagram, each block represents 8 bits (1 byte) of data. Our memory is byte-addressed, meaning that there is one address for each byte. Compare this to word-addressed, which means that there is one address for each word. The Geometry of Caches Main Cache Main Index Offset Memory Number Memory CPU Number 3 2 1 0 ... 3 2 1 0 Cache 0 0 6 7 6 5 4 1 5 11 10 9 8 1 Tag bits Index bits Offset bits Total 4 15 14 13 12 29 1 2 32 3 19 18 17 16 2 2 23 22 21 20 1word=4 bytes=32bits Index bits=log (Number of index rows) 1 27 26 25 24 2 3 Offset bits=log2(Number of offsets columns) 0 31 30 29 28 1. Direct mapped caches 1. How many bytes of data can our cache hold? 8 bytes How many words? 2 words 2. Fill in the “Tag bits, Index bits, Offset bits” with the correct T:I:O breakdown according to the diagram. 3. Let’s say we have a 8192KiB cache with an 128B block siZe, what is the tag, index, and offset of 0xFEEDF00D? FE ED F0 0D 1111 1110 1110 1101 1111 0000 0000 1101 Tag: 1 1111 1101 (0x1FD) Index: 1101 1011 1110 0000 (0xDBE0) Offset: 000 1101 (0x0D) 4. Fill in the table below. Assume we have a write-through cache, so the number of bits per row includes only the cache data, the tag, and the valid bit.
    [Show full text]
  • Memory We Have Already Mentioned That Digital Computer Works on Stored Programmed Concept Introduced by Von Neumann
    www.getmyuni.com Memory We have already mentioned that digital computer works on stored programmed concept introduced by Von Neumann. We use memory to store the information, which includes both program and data. Due to several reasons, we have different kind of memories. We use different kind of memory at different level. The memory of computer is broadly categories into two categories: . Internal and . external Internal memory is used by CPU to perform task and external memory is used to store bulk information, which includes large software and data. Memory is used to store the information in digital form. The memory hierarchy is given by: . Register . Cache Memory . Main Memory . Magnetic Disk . Removable media (Magnetic tape) Register: This is a part of Central Processor Unit, so they reside inside the CPU. The information from main memory is brought to CPU and keep the information in register. Due to space and cost constraints, we have got a limited number of registers in a CPU. These are basically faster devices. Cache Memory: Cache memory is a storage device placed in between CPU and main memory. These are semiconductor memories. These are basically fast memory device, faster than main memory. We cannot have a big volume of cache memory due to its higher cost and some constraints of the CPU. Due to higher cost we cannot replace the whole main memory by faster memory. Generally, the most recently used information is kept in the cache memory. It is brought from the main memory and placed in the cache memory. Now days, we get CPU with internal cache.
    [Show full text]
  • VLSI Memory Lecture-Nahas-181025.Pdf
    Introduction to CMOS VLSI Design Semiconductor Memory Harris and Weste, Chapter 12 25 October 2018 J. J. Nahas and P. M. Kogge Modified from slides by Jay Brockman 2008 [Including slides from Harris & Weste, Ed 4, Adapted from Mary Jane Irwin and Vijay Narananan, CSE Penn State adaptation of Rabaey’s Digital Integrated Circuits, ©2002, J. Rabaey et al.] Semiconductor Memory Slide 1 Outline Memory features and comparisons Generic Memory Architecture – Architecture Overview – Row and Column Decoders – Redundancy and Error Correction Static Random Access Memory (SRAM) Dynamic Random Access Memory (DRAM) Flash (EEPROM) Memory Other Memory Types Semiconductor MemoryCMOS VLSI Design Slide 2 1 Memory Features and Comparisons Semiconductor MemoryCMOS VLSI Design Slide 3 Memory Characteristics Read/Write Attributes – Read-Only Memory (ROM): Programmed at manufacture • Being phased out of use in favor of Flash – Read-Write Memory: Can change value dynamically • SRAM, DRAM – Read-Mostly: Can write, but much more slowly than read • EEPROM (Electrically Eraseable, Programable Read Only Memory) (pronounced “Double E Prom”) • Flash (A form of EEPROM) Volatility: sensitivity to losing power – Volatile: loses contents when power turned off – Non-volatile: does not lose contents Semiconductor MemoryCMOS VLSI Design Slide 4 2 Memory Characteristics Addressability – Random-Access: provide address to access a “word” of data • No correlation between successive addresses – Block oriented: read and write large blocks of data at a time – Content-addressable:
    [Show full text]