Flashy Prefetching for High-Performance Flash Drives

Total Page:16

File Type:pdf, Size:1020Kb

Flashy Prefetching for High-Performance Flash Drives Flashy Prefetching for High-Performance Flash Drives Ahsen J. Uppal, Ron C. Chiang, and H. Howie Huang Department of Electrical and Computer Engineering George Washington University fauppal, rclc, [email protected] Abstract—While hard drives hold on to the capacity advantage, seek penalties, limited HDD bandwidth, and limited system flash-based solid-state drives (SSD) with high bandwidth and RAM), with one notable exception [36]. low latency have become good alternatives for I/O-intensive For flash-based SSDs, we believe that aggressive prefetch- applications. Traditional data prefetching has been primarily designed to improve I/O performance on hard drives. The same ing could potentially expedite data requests for many ap- techniques, if applied unchanged on flash drives, are likely to plications. However, as we will demonstrate shortly, simply either fail to fully utilize SSDs, or interfere with application I/O prefetching as much data as possible does not provide the requests, both of which could result in undesirable application desired benefits for three main reasons. First, data prefetching performance. In this work, we demonstrate that data prefetching, on faster devices such as SSDs, if uncontrolled, will take when effectively harnessing the high performance of SSDs, can provide significant performance benefits for a wide range of shared I/O bandwidth from existing data accesses (more easily data-intensive applications. The new technique, flashy prefetching, than on slower hard drives). As a side effect, useful cached consists of accurate prediction of application needs in runtime data may become evicted while main memory would be filled and adaptive feedback-directed prefetching that scales with with mispredicted (and unneeded) data while applications were application needs, while being considerate to underlying storage waiting for useful data. devices. We have implemented a real system in Linux and evaluated it on four different SSDs. The results show 65-70% Second, not every device is the same, and this is especially prefetching accuracy and an average 20% speedup on LFS, web true for SSDs. The performance of an SSD can vary depending search engine traces, BLAST, and TPC-H like benchmarks across on flash type (SLC/MLC), internal organization, memory various storage drives. management, etc. A prefetching algorithm, while reasonably aggressive for a faster drive, could become too aggressive for I. INTRODUCTION another drive, slowing down normal execution. Last, not every The spectrum of storage devices has expanded drastically in application is the same – two applications often possess differ- the last several years, thanks to the emergence of solid-state ent I/O requirements. A single application can also go through drives (SSDs) that are built upon NAND flash memory [6], multiple stages, each of which has different I/O requirements. [13]. As scientific and enterprise data usage continues to grow Clearly, care should be taken to avoid adverse effects from exponentially, new storage systems that leverage both high both too-conservative and too-aggressive prefetching. performance from SSDs and large capacity from hard drives In this work, we propose the technique of flashy prefetch- (HDDs) will likely be in high demand to reduce the I/O ing1for emerging flash devices, which is aware of the runtime performance gap. While for many data-intensive applications environment and can adapt to the changing requirements of moving, ”hot” data from HDDs to SSDs (disk swapping and both devices and applications as well as its own run-time data migration) can easily bring good speedups, in this paper performance. we aim to achieve additional performance benefits from SSDs The salient features of flashy prefetching include not only beyond the simple practice of disk replacement. taking advantage of the high bandwidth and low latency of Data prefetching [37], [19] is one of, if not the most, SSDs, but also providing inherent support for parallel data widely-used techniques to reduce access latency, because it accesses and feedback-controlled aggressiveness. To demon- can load the data that are likely to soon be accessed from strate its feasibility and benefits, we have implemented a real storage devices into main memory. Traditional prefetching system called prefetchd in Linux that dynamically controls techniques have been focused on rotational hard drives and its prefetching aggressiveness at runtime to maximize per- are conservative with the amount of data prefetched for good formance benefits, by making good tradeoffs between data reasons – because data prefetching consumes shared system prefetching and resource consumption. Note that we use flashy resources, it is likely that aggressive data prefetching would in- prefetching and prefetchd interchangeably in this paper. We terfere with normal access and subsequently hinder application evaluate prefetchd on four different SSDs with a wide range performance. As a result, current techniques often leverage the of data-intensive applications and benchmarks. The prototype low cost of sequential access on hard drives to read data that resides on the same and nearby tracks. Aggressive prefetching 1”Flashy: momentarily dazzling.” (source: Merriam-Webster) Flashy has been considered too risky by many researchers (given long prefetching aims to effectively harness ”flashy”, high-performance SSDs. 978-1-4673-1747-4/12/$31.00 c 2013 IEEE TABLE I achieves average 20% speedups on LFS, web search engine SPECIFICATION AND MEASURED PERFORMANCE traces, BLAST, and TPC-H like benchmarks across various SSD1 [34] SSD2 [35] SSD3 [23] SSD4 [24] storage drives, which we believe largely comes from the 65- Capacity 120 GB 120 GB 80 GB 120 GB 70% prefetching accuracy. Flash MLC MLC MLC MLC The main contributions of this paper are threefold: Type 34nm 34nm 25nm Controller Indilinx SandForce Intel Marvell • We conduct a comprehensive study of the effects of Read BW 250 MB/s 285 MB/s 250 MB/s 450 MB/s conservative and aggressive prefetching in the context (Bandwidth) (max) (max) (seq) (seq) Write BW 180 MB/s 275 MB/s 100 MB/s 210 MB/s of heterogeneous devices and applications. The results (max) (max) (seq) (seq) show that adaptive prefetching is essential for taking Latency 100 ms 100 ms 65 ms 65 ms advantage of high-performance storage devices like solid- read read state drives. Measured 170 MB/s 170 MB/s 265 MB/s 215 MB/s Read BW • We design the architecture of flashy prefetching such that Measured 180 MB/s 203 MB/s 81 MB/s 212 MB/s it self-tunes to prefetch data in a manner that matches Write BW application needs without being so aggressive that useful pages are evicted from the cache. Measuring performance SSDs from two manufacturers (roughly covering two recent metrics in real-time and adjusting the aggressiveness generations of SSDs): OCZ Vertex (SSD1) [34] and Vertex2 accordingly significantly improves the effectiveness of (SSD2) [35], and Intel X-25M (SSD3) [23] and 510 (SSD4) this approach. [24]. We also use a Samsung Spinpoint M7 (HDD) hard drive [39]. If we look at the specifications in the top half of Table I, • We develop a Linux-based prototype, prefetchd, that the specification numbers for SSDs are close, except for the monitors application read requests, predicts which pages Intel 510 SSD (which comes with the SATA III support, but are likely to be read in the near future, loads those pages is limited by the SATA II interface in our test machines). into the system page cache while attempting to not evict However, the differences between different SSDs tend to other useful pages, monitors its success rate in time and be subtle, mostly in architectural designs. Note that the same across pages, and adjusts its aggressiveness accordingly. manufacturer may choose to adopt different controllers across Note that data prefetching on SSDs has drawn a lot of two models, which coincidentally was the case for the SSDs interest. For example, [46], [9] show that prefetching can be chosen in this study. As shown in the bottom half in Table I, used to improve energy efficiency. Our positive results also when measured under Linux, the four SSDs clearly have demonstrate the performance potential of data prefetching. higher bandwidth than the hard drive (measured read band- Another notable work, FAST [25], focuses on shortening width at about 90 MB/s), that is, the four SSDs outperform application launch times and utilizes prefetching on SSDs the hard drive by 189%, 189%, 294%, and 239%, respectively. for the quick start of various applications. Our approach ex- More importantly, the four SSDs differ noticeably, especially, pands on previous work along multiple dimensions, including their measured write bandwidths range from 80 MB/s and 200 employing a feedback-driven control mechanism, handling MB/s. multiple simultaneous requests across processes, and requiring no application modifications. With a wider range of data- 7000 intensive applications in mind, the proposed prefetchd aims 6000 to improve the overall performance of generic applications. 5000 The rest of the paper is organized as follows: Section 4000 II presents the challenges the prefetching must address and Section III describes our design principles to address those 3000 challenges. Section IV presents the architecture of prefetchd 2000 Throughput (IOPS) and describes each individual component. Section V discusses 1000 the implementation in detail. The evaluation is presented in Section VI and related work is discussed in Section VII. We 0 conclude in Section VIII. lfs ws1 ws2 db1 db2 db3 db4 db5 db6 db7 Benchmarks II. THREE CHALLENGES Fig. 1. Application I/O Throughputs A. Challenge #1: SSDs are different SSDs are clearly different from HDDs in many ways. To B. Challenge #2: Applications are different name a few: no seek latency, excellent random read and write Although data-intensive applications are in dire need of performance, inherent support for parallel I/O, expensive small high-performance data access, they tend to have different writes, and limited erase cycles.
Recommended publications
  • Partitioner Och Filsystem 2
    Partitioner och filsystem 2 File systems FAT Unix-like NTFS Vad är ett filsystem? • Datorer behöver en metod för att lagra och hämta data… • Referensmodell för filsystem (Carrier) – Filsystem kategori • Layout och storleksinformation – Innehålls kategori • Kluster och block – data enheter – Metadata kategori • Tidsinformation, storlek, access kontroll • Adresser till allokerade data enheter – Filnamn kategori • Oftast ihop-kopplad med metadata – Applikations kategori • Quota • Journaler • De modernaste påminner mycket om relations databaser Windows • NTFS (New Technology File System) – 6 versioner finns, de nyaste är v3.0 (Windows 2000) och v3.1 (XP, 2003, Vista, 2008, 7), kallas även 5.0, 5.1, 5.2, 6.0 och 6.1 (efter OS version) – Stöd för unicode, säkerhet, mm. - är mycket mer komplext än FAT! – http://en.wikipedia.org/wiki/Ntfs • FAT 12/16/32, VFAT (långa filnamn i Win95) – Används fortfarande men är inte effektivt för större lagringskapaciteter (klusterstorleken) – Långsammare access än NTFS • Windows Future Storage (WinFS) inställt projekt, enligt rykten var det en SQL-databas som ligger ovanpå ett NTFS filsystem – Läs mer på: http://www.ntfs.com/ – Och: http://en.wikipedia.org/wiki/WinFS FAT12, 16 och 32 • FAT12, finns på floppy diskar – Begränsad lagringskapacitet – Designat för MS-DOS 1.0 • FAT16, var designat för större diskar – Äldre OS använde detta • MS-DOS 3.0, Win95 OSR1, NT 3.5 och NT 4.0 – Max diskstorlek 2 GB • FAT32 kom när diskar större än 2GB kom – Vissa äldre och alla nya OS kan använda FAT32 • Windows 98/Me/2000/XP/2003/Vista/7 och 2008 • Begränsningar med FAT32 – Största formaterabara volymen är 32GB (större volymer kan dock användas, < 16 TiB) – Begränsade features vad gäller komprimering, kryptering, säkerhet och hastighet jämfört mot NTFS • http://en.wikipedia.org/wiki/FAT_file_system exFAT • exFAT (Extended File Allocation Table, a.k.a.
    [Show full text]
  • PC Gamers Win Big
    CASE STUDY Intel® Solid-State Drives Performance, Storage and the Customer Experience PC Gamers Win Big Intel® Solid-State Drives (SSDs) deliver the ultimate gaming experience, providing dramatic visual and runtime improvements Intel® Solid-State Drives (SSDs) represent a revolutionary breakthrough, delivering a giant leap in storage performance. Designed to satisfy the most demanding gamers, media creators, and technology enthusiasts, Intel® SSDs bring a high level of performance and reliability to notebook and desktop PC storage. Faster load times and improved graphics performance such as increased detail in textures, higher resolution geometry, smoother animation, and more characters on the screen make for a better gaming experience. Developers are now taking advantage of these features in their new game designs. SCREAMING LOAD TiMES AND SMOOTH GRAPHICS With no moving parts, high reliability, and a longer life span than traditional hard drives, Intel Solid-State Drives (SSDs) dramatically improve the computer gaming experience. Load times are substantially faster. When compared with Western Digital VelociRaptor* 10K hard disk drives (HDDs), gamers experienced up to 78 percent load time improvements using Intel SSDs. Graphics are smooth and uninterrupted, even at the highest graphics settings. To see the performance difference in a head-to- head video comparing the Intel® X25-M SATA SSD with a 10,000 RPM HDD, go to www.intelssdgaming.com. When comparing frame-to-frame coherency with the Western Digital VelociRaptor 10K HDD, the Intel X25-M responds with zero hitching while the WD VelociRaptor shows hitching seven percent of the time. This means gamers experience smoother visual transitions with Intel SSDs.
    [Show full text]
  • Enhance Your Customer's Vista System with Datatraveler Readyflash
    kingston.com/windowsvista Enhance your customer’s Vista system with DataTraveler ReadyFlash For more information, Complement your Vista™ sales with a Kingston® for Windows ReadyBoost with at least 512 MB of please contact your dedicated DataTraveler® ReadyFlash USB Flash drive enhanced available space. Kingston representative: for Windows® ReadyBoost™. Craig McQueenie – Central Can I enable ReadyBoost with any USB (416) 315-9097 Ideal for systems with limited memory expansion Flash drive? [email protected] capabilities, Windows ReadyBoost allows a user to No, Vista will only offer the option of using Daniel Ouimet – Quebec utilize memory outside of system memory module ReadyBoost on USB Flash drives that have met both (514) 244-4901 upgrades. For instance, if a notebook has a maximum a specific performance and a capacity threshold. [email protected] memory capacity of 1 GB installed, Windows Patrick Jan – West ReadyBoost allows for usage of a Flash memory Will Kingston offer additional Flash memory (604) 961-8038 device, such as DataTraveler ReadyFlash as a more devices for ReadyBoost? [email protected] responsive storage medium for ReadyBoost cache. Yes, Kingston will offer additional Flash products that Royal Taylor – Central (416) 617-6714 Kingston Part Number: DTR/1GB are enhanced for Windows ReadyBoost. Please contact [email protected] your Kingston representative for further details. Description: 1-GB USB Flash drive enhanced for Windows ReadyBoost. Can I use more than one ReadyBoost enabled device at a time? Windows Vista allows for only one device enabled FAQ: for ReadyBoost at a time. What is ReadyBoost? ReadyBoost is a Microsoft Windows Vista feature How much memory should I allocate for which may improve system performance by ReadyBoost? allowing a compatible USB drive to be configured If you have selected to use ReadyBoost, you can with a ReadyBoost cache.
    [Show full text]
  • Azor: Using Two-Level Block Selection to Improve SSD-Based I/O Caches
    Azor: Using Two-level Block Selection to Improve SSD-based I/O caches Yannis Klonatos∗†, Thanos Makatos∗, Manolis Marazakis∗, Michail D. Flouris∗, and Angelos Bilas∗† ∗Foundation for Research and Technology - Hellas (FORTH), Institute of Computer Science (ICS) 100 N. Plastira Ave., Vassilika Vouton, Heraklion, GR-70013, Greece †Department of Computer Science, University of Crete, P.O. Box 2208, Heraklion, GR 71409, Greece. {klonatos, mcatos, maraz, flouris, bilas}@ics.forth.gr Abstract—Flash-based solid state drives (SSDs) exhibit poten- approaches that are application-specific and require application tial for solving I/O bottlenecks by offering superior performance knowledge, intervention, and tuning, Azor transparently and over hard disks for several workloads. In this work we design dynamically places data blocks in the SSD cache as they flow Azor, an SSD-based I/O cache that operates at the block-level and is transparent to existing applications, such as databases. Our in the I/O path between main memory and HDDs. In this work, design provides various choices for associativity, write policies we investigate the following problems: and cache line size, while maintaining a high degree of I/O concurrency. Our main contribution is that we explore differ- TABLE I entiation of HDD blocks according to their expected importance HDD AND SSD PERFORMANCE METRICS. on system performance. We design and analyze a two-level block selection scheme that dynamically differentiates HDD blocks, and SSD HDD selectively places them in the limited space of the SSD cache. Price/capacity ($/GB) $3 $0.3 We implement Azor in the Linux kernel and evaluate its Response time (ms) 0.17 12.6 effectiveness experimentally using a server-type platform and Throughput (R/W) (MB/s) 277/202 100/90 large problem sizes with three I/O intensive workloads: TPC-H, IOPS (R/W) 30,000/3,500 150/150 SPECsfs2008, and Hammerora.
    [Show full text]
  • SPDK FTL & Marvell OCSSD for Noisy Neighbor Problem
    SPDK FTL & Marvell OCSSD for Noisy Neighbor Problem David Recker Marketing VP Circuit Blvd., Inc. April 2019 4/17/2019 © 2019 Circuit Blvd., Inc. 1 Industry Who We Are Enterprise/Cloud Database and Storage Year Founded Sunnyvale, CA, U.S.A. 2017 Mission We develop next gen database/storage systems leveraging expertise in memory semiconductor, solid-state storage system, and operating systems Open Source Contributions • Linux LightNVM, OCSSD 2.0 specification, OpenSSD FPGA platform • SPDK (since SPDK v17.10) • RocksDB 4/17/2019 © 2019 Circuit Blvd., Inc. 2 OCSSD with SPDK FTL • SPDK FTL on Marvell’s OCSSD Platform • We have been evaluating SPDK FTL on Marvell's SSD SoC platform since Jan ’19 • SPDK (Flash Translation Layer) FTL: The Flash Translation Layer library provides block device access on top of non-block SSDs implementing Open Channel interface. It handles the logical to physical address mapping, responds to the asynchronous media management events, and manages the defragmentation process* • Measured various performance metrics of initial prototype and demonstrate how SPDK OCSSDs can solve the noisy neighbor problem in multi-tenant environments • Share experimental data based on our current implementation (both SPDK FTL and Marvell’s controller being continuously improved) • (Demo) SPDK Driven OCSSD Comparison (Isolation vs Non-Isolation) • Demo table outside (please feel free to drop by for further questions) * SPDK FTL definition: https://spdk.io/doc/ftl.html 4/17/2019 © 2019 Circuit Blvd., Inc. 3 Hardware Setup • SuperMicro X11DPG • 2 * Xeon Scalable Gold 6126 2.6 Ghz (12 cores) • hyperthreading disabled OCSSD1 • 8 * 32 GB DIMM 2666 MT/s • 2 * OCSSD 2.0 OCSSD2 • Marvell 88SS1098 controller • PCIe Gen3x4 slot to each CPU package • nvme id-ns • LBADS=12 (4KiB), MS=0 • ocssd geometry • 8 grp (3), 8 pu (3), 1478 chk (11), 6144 lbk (13) • () means bit length in LBAF • ws_opt=24 (96KiB) CPU1 • 3D TLC NAND CPU2 • write unit: 96KiB (one shot program) • read unit: 32KiB 4/17/2019 © 2019 Circuit Blvd., Inc.
    [Show full text]
  • Information Hiding at SSD NAND Flash Memory Physical Layer
    SECURWARE 2014 : The Eighth International Conference on Emerging Security Information, Systems and Technologies DeadDrop-in-a-Flash: Information Hiding at SSD NAND Flash Memory Physical Layer Avinash Srinivasan and Jie Wu Panneer Santhalingam and Jeffrey Zamanski Temple University George Mason University Computer and Information Sciences Volgenau School of Engineering Philadelphia, USA Fairfax, USA Email: [avinash, jiewu]@temple.edu Email: [psanthal, jzamansk]@gmu.edu Abstract—The research presented in this paper, to of logical blocks to physical flash memory is controlled by the best of our knowledge, is the first attempt at the FTL on SSDs, this cannot be used for IH on SSDs. Our information hiding (IH) at the physical layer of a Solid proposed solution is 100% filesystem and OS-independent, State Drive (SSD) NAND flash memory. SSDs, like providing a lot of flexibility in implementation. Note that HDDs, require a mapping between the Logical Block throughout this paper, the term “physical layer” refers Addressing (LB) and physical media. However, the to the physical NAND flash memory of the SSD, and mapping on SSDs is significantly more complex and is handled by the Flash Translation Layer (FTL). FTL readers should not confuse it with the Open Systems is implemented via a proprietary firmware and serves Interconnection (OSI) model physical layer. to both protect the NAND chips from physical access Traditional HDDs, since their advent more than 50 as well as mediate the data exchange between the years ago, have had the least advancement among storage logical and the physical disk. On the other hand, the hardware, excluding their storage densities.
    [Show full text]
  • INFORMATION TECHNOLOGY CONCEPTS-OPEN - REGIONAL 2019 Page 1 of 8
    INFORMATION TECHNOLOGY CONCEPTS-OPEN - REGIONAL 2019 Page 1 of 8 INFORMATION TECHNOLOGY CONCEPTS (391) —OPEN EVENT— REGIONAL – 2019 DO NOT WRITE ON TEST BOOKLET TOTAL POINTS _________ (100 points) Failure to adhere to any of the following rules will result in disqualification: 1. Contestant must hand in this test booklet and all printouts. Failure to do so will result in disqualification. 2. No equipment, supplies, or materials other than those specified for this event are allowed in the testing area. No previous BPA tests and/or sample tests or facsimile (handwritten, photocopied, or keyed) are allowed in the testing area. 3. Electronic devices will be monitored according to ACT standards. No more than sixty (60) minutes testing time Property of Business Professionals of America. May be reproduced only for use in the Business Professionals of America Workplace Skills Assessment Program competition. INFORMATION TECHNOLOGY CONCEPTS-OPEN - REGIONAL 2019 Page 2 of 8 MULTIPLE CHOICE Identify the choice that best completes the statement or answers the question. Mark A if the statement is true. Mark B if the statement is false. 1. Which of the following appears on the right side of any Windows 8 screen when you move your pointer to a right corner? A. Live tile B. Memory Manager C. Charms bar D. System tray 2. Which element of the Windows 7 GUI gives windows a glassy appearance, but also consumes more hardware resources? A. Control panel B. Aero user interface C. Charms interface D. Logic interface 3. The top of a top-down hierarchical structure of subdirectories is called which of the following? A.
    [Show full text]
  • Micron: NAND Flash Architecture and Specification Trends
    NAND Flash Architecture and Specification Trends Michael Abraham ([email protected]) Applications Engineering Manager Micron Technology, Inc. Santa Clara, CA USA August 2009 1 Abstract As NAND Flash continues to shrink, page sizes, block sizes, and ECC requirements are increasing while data retention, endurance, and performance are decreasing. These changes impact systems including random write performance and more. Learn how to prepare for these changes and counteract some of them through improved block management techniques and system design. This presentation also discusses some of the tradeoff myths – for example, the myth that you can directly trade ECC for endurance Santa Clara, CA USA August 2009 2 NAND Flash: Shrinking Faster Than Moore’s Law 200 100 Logic 80 DRAM on (nm) ii 60 NAND Resolut 40 Micron 32Gb NAND (34nm) 2000 2001 2002 2003 2004 2005 2006 2007 2008 2009 2010 2011 2012 Semiconductor International, 1/1/2007 Santa Clara, CA USA August 2009 3 Memory Organization Trends Over time, NAND block size is increasing. • Larger page sizes increase sequential throughput. • More pages per block reduce die size. 4,194,304 1,048,576 262,144 65,536 16,384 4,096 1, 024 256 64 16 Block size (B) Data Bytes per Page Pages per Block Santa Clara, CA USA August 2009 4 Consumer-grade NAND Flash: Endurance and ECC Trends Process shrinks lead to less electrons ppgger floating gate. ECC used to improve data retention and endurance. To adjust for increasing RBERs, ECC is increasing exponentially to achieve equivalent UBERs. For consumer applications, endurance becomes less important as density increases.
    [Show full text]
  • SSD ABC Guide – OCZ Forum Ii © 2014 OCZ Storage Solutions
    SSD ABC Guide – OCZ Forum ii © 2014 OCZ Storage Solutions SSD ABC Guide – OCZ Forum Contents 1 - 8 .......................................................................................... Identifying your SSD ..................................................................................................................................................... 1 Products ........................................................................................................................................................................ 2 Unpacking and Handling your SSD .............................................................................................................................. 3 Fitting your SSD ............................................................................................................................................................ 4 Desktop ...................................................................................................................................................................................... 4 Laptop/Notebook ........................................................................................................................................................................ 4 Powering On and Self Testing (POST) your SSD ......................................................................................................... 4 Detecting your SSD ......................................................................................................................................................
    [Show full text]
  • Flash-Aware Database Management Systems Hardock, Sergej (2020)
    Flash-aware Database Management Systems Hardock, Sergej (2020) DOI (TUprints): https://doi.org/10.25534/tuprints-00014476 License: CC-BY-SA 4.0 International - Creative Commons, Attribution Share-alike Publication type: Ph.D. Thesis Division: 20 Department of Computer Science Original source: https://tuprints.ulb.tu-darmstadt.de/14476 Flash-aware Database Management Systems Zur Erlangung des akademischen Grades Doktor-Ingenieur (Dr.-Ing.) Genehmigte Dissertation von Sergej Hardock aus Kiew, Ukraine Tag der Einreichung: 30.7.2020, Tag der Prüfung: 4.11.2020 1. Gutachten: Prof. Dr. Carsten Binnig 2. Gutachten: Prof. Dr.-Ing. Ilia Petrov 3. Gutachten: Alejandro Buchmann, Ph.D. Darmstadt Computer Science Department Technical University of Darmstadt Data Management Lab Flash-aware Database Management Systems Accepted doctoral thesis by Sergej Hardock 1. Review: Prof. Dr. Carsten Binnig 2. Review: Prof. Dr.-Ing. Ilia Petrov 3. Review: Alejandro Buchmann, Ph.D. Date of submission: 30.7.2020 Date of thesis defense: 4.11.2020 Darmstadt Bitte zitieren Sie dieses Dokument als: URN: urn:nbn:de:tuda-tuprints-144769 URL: http://tuprints.ulb.tu-darmstadt.de/id/eprint/14476 Dieses Dokument wird bereitgestellt von tuprints, E-Publishing-Service der TU Darmstadt http://tuprints.ulb.tu-darmstadt.de [email protected] Die Veröffentlichung steht unter folgender Creative Commons Lizenz: Namensnennung - Weitergabe unter gleichen Bedingungen 4.0 International https://creativecommons.org/licenses/by-sa/4.0 This dissertation is dedicated to my parents who encouraged me to follow my dreams. Erklärungen laut Promotionsordnung §8 Abs. 1 lit. c PromO Ich versichere hiermit, dass die elektronische Version meiner Dissertation mit der schriftli- chen Version übereinstimmt.
    [Show full text]
  • Windows Internals, Sixth Edition, Part 2
    spine = 1.2” Part 2 About the Authors Mark Russinovich is a Technical Fellow in ® the Windows Azure™ group at Microsoft. Windows Internals He is coauthor of Windows Sysinternals SIXTH EDITION Administrator’s Reference, co-creator of the Sysinternals tools available from Microsoft Windows ® The definitive guide—fully updated for Windows 7 TechNet, and coauthor of the Windows Internals and Windows Server 2008 R2 book series. Delve inside Windows architecture and internals—and see how core David A. Solomon is coauthor of the Windows Internals book series and has taught components work behind the scenes. Led by a team of internationally his Windows internals class to thousands of renowned internals experts, this classic guide has been fully updated Windows developers and IT professionals worldwide, SIXTH for Windows 7 and Windows Server® 2008 R2—and now presents its including Microsoft staff. He is a regular speaker 6EDITION coverage in two volumes. at Microsoft conferences, including TechNet As always, you get critical, insider perspectives on how Windows and PDC. operates. And through hands-on experiments, you’ll experience its Alex Ionescu is a chief software architect and internal behavior firsthand—knowledge you can apply to improve consultant expert in low-level system software, application design, debugging, system performance, and support. kernel development, security training, and Internals reverse engineering. He teaches Windows internals courses with David Solomon, and is ® In Part 2, you will: active in the security research community.
    [Show full text]
  • Arxiv:1901.01161V1 [Cs.CR] 4 Jan 2019
    Page Cache Attacks Daniel Gruss1, Erik Kraft1, Trishita Tiwari2, Michael Schwarz1, Ari Trachtenberg2, Jason Hennessey3, Alex Ionescu4, Anders Fogh5 1 Graz University of Technology, 2 Boston University, 3 NetApp, 4 CrowdStrike, 5 Intel Corporation Abstract last twenty years [3, 40, 53]. Osvik et al. [51] showed that an attacker can observe the cache state at the granularity of We present a new hardware-agnostic side-channel attack that a cache set using Prime+Probe, and later Yarom et al. [77] targets one of the most fundamental software caches in mod- showed this with cache line granularity using Flush+Reload. ern computer systems: the operating system page cache. The While different cache attacks have different use cases, the page cache is a pure software cache that contains all disk- accuracy of Flush+Reload remains unrivaled. backed pages, including program binaries, shared libraries, Indeed, virtually all Flush+Reload attacks target pages in and other files, and our attacks thus work across cores and the so-called page cache [30, 33, 34, 35, 42, 77]. The page CPUs. Our side-channel permits unprivileged monitoring of cache is a pure software cache implemented in all major op- some memory accesses of other processes, with a spatial res- erating systems today, and it contains virtually all pages in olution of 4 kB and a temporal resolution of 2 µs on Linux use. Pages that contain data accessible to multiple programs, (restricted to 6:7 measurements per second) and 466 ns on such as disk-backed pages (e.g., program binaries, shared li- Windows (restricted to 223 measurements per second); this braries, other files, etc.), are shared among all processes re- is roughly the same order of magnitude as the current state- gardless of privilege and permission boundaries [24].
    [Show full text]