An Effective DRAM Cache Architecture for Scale-Out Servers

An Effective DRAM Cache Architecture for Scale-Out Servers

This technical report is a longer version of "Fat Caches for Scale-Out Servers" which appears in IEEE Micro. An Effective DRAM Cache Architecture for Scale-Out Servers Stavros Volos Djordje Jevdjic Babak Falsafi Boris Grot Microsoft Research University of Washington EcoCloud, EPFL University of Edinburgh ABSTRACT that maximize compute density. In response, processor ven- Scale-out workloads are characterized by in-memory dors have turned to customized architectures for datacenters. datasets, and consequently massive memory footprints. Due Following a considerable amount of research, identifying to the abundance of request-level parallelism found in these the requirements of scale-out workloads and indicating that workloads, recent research advocates for manycore architec- these workloads benefit from thread-level parallelism and tures to maximize throughput while maintaining quality of high core counts [27, 48, 61], industry has started employ- service. On-die stacked DRAM caches have been proposed ing specialized manycore processors (e.g., Cavium Thun- to provide the required bandwidth for manycore servers derX [32] and EZchip Tile-MX [33]) due to the substantial through caching of secondary data working sets. How- performance and TCO advantages offered by specialization. ever, the disparity between provided capacity and hot dataset Memory systems in scale-out servers are of paramount working set sizes — resulting from power-law dataset ac- importance as they have to sustain the vast bandwidth de- cess distributions — precludes their effective deployment in mands of manycore CMPs [45, 61]. Recent advances in servers, calling for high-capacity cache architectures. on-die stacked DRAM technology [10, 59] can eliminate the In this work, we find that while emerging high-bandwidth bandwidth bottleneck that plagues conventional DRAM. As memory technology falls short of providing enough capacity this technology is capacity-limited due to thermal constraints to serve as system memory, it is a great substrate for high- [22], prior research advocates for using it as a cache [44, 45, capacity caches. We also find the long cache residency peri- 46, 60, 81] to provide access to secondary data working sets. ods enabled by high-capacity caches uncover significant spa- Our analysis shows that on-die stacked DRAM caches are tial locality across objects. Based on our findings, we intro- unattractive for scale-out servers. We find that memory ac- duce Scale-Out Cache (soCache), a distributed cache com- cesses follow power-law distributions — similarly to enter- posed of multiple high-bandwidth memory modules. Each prise server workloads [62] — with a hot portion of memory soCache module uses a page-based organization that opti- (~10%) accounting for the bulk of accesses (65-95%). Thus, mizes for spatial locality while minimizing tag storage re- while the vast working sets of scale-out workloads are cache- quirements. By storing the tags in the logic die (in SRAM) able, high-capacity caches (10s of GB) are required given of the high-bandwidth memory modules, soCache avoids the main memory sizes trending toward 100s of GB. The re- prohibitive complexity of in-DRAM metadata in state-of- quired cache capacities greatly exceed those of low-capacity the-art DRAM caches. In 14nm technology, soCache re- caches, including on-die stacked DRAM caches. duces system energy by 1.4-4.4x and improves throughput This work seeks to develop a scalable, high-capacity, by 28-44% over state-of-the-art memory systems. and high-bandwidth memory system for scale-out servers by leveraging emerging high-bandwidth memory modules as 1. INTRODUCTION a high-capacity cache. High-bandwidth interconnect tech- nologies allow for connecting the processor to multiple high- Scale-out datacenters host a variety of data-intensive ser- bandwidth memory modules via a silicon interposer [50] vices, such as search and social connectivity. To concur- forming an on-package high-capacity cache, or high-speed rently support billions of users, latency-sensitive online ser- serial links [75] forming an off-package high-capacity cache. vices rely on large amounts of memory to avoid disk ac- We draw insights on design of high-capacity caches by cesses [11, 73]. Likewise, analytic engines creating user- characterizing their access patterns, finding that they exhibit specific content, such as advertisements and recommenda- predominantly coarse access granularity, with 93% of all ac- tions, rely on massive memory capacity [18] as they need cesses going to memory regions with high access density.1 to plow through enormous amounts of data within 100s In contrast, only 68% of accesses in low-capacity caches are of milliseconds. While today’s datacenters rely solely on coarse-grained. The reason why high-capacity caches ex- DRAM, the emergence of fast NVRAM will further broaden hibit higher access density is due to the longer residency in-memory computing. The ever-growing popularity of the times of data in the cache, which increases the likelihood in-memory computing paradigm leads to datacenter deploy- ments in which memory accounts for a big share of the dat- 1 A high access density region is one in which most (over 75%) of acenter’s total cost of ownership (TCO) [8, 52, 65]. cache blocks comprising the region are accessed within the cache Optimizing for datacenter’s TCO calls for architectures lifetime of the first block to be accessed within the region [93]. for unrelated spatially-proximate objects (e.g., adjacent hash Table 1: Requirements of one scale-out server. bucket headers) to be accessed before an eviction. Based on the insights of the characterization, we intro- Year Processor Memory System duce MeSSOS, a Memory System architecture for Scale- Cores Bandwidth Bandwidth Capacity Out Servers. MeSSOS employs a set of off-package high- 2015 96 115 GB/s 288 GB/s 384 GB bandwidth memory modules as a high-capacity hardware- managed scale-out cache (soCache). Each soCache module 2018 180 216 GB/s 540 GB/s 720 GB is backed by a number of conventional DIMMs. Together, 2021 320 384 GB/s 960 GB/s 1280 GB an soCache module and its associated DIMMs represent a MeSSOS building block. Growing (cache and memory) ca- pacity and bandwidth to accommodate more cores and larger 2.1 Scale-Out Server Requirements datasets is simply a matter of adding more building blocks. The architecture of an soCache module is informed by the Large-scale online services distribute and replicate their memory access behavior of scale-out workloads. Due to the datasets across many servers to ensure in-memory process- prevalence of coarse-grained accesses and long cache resi- ing and meet tight tail latency requirements. In response, dency times stemming from skewed access distributions, so- processor and system vendors resort to manycore processors Cache employs a page-based organization. Tags are stored [32, 89, 92, 95] and buffer-on-board chips [17, 40, 92] to in the logic die of each soCache module, taking advantage boost per-server throughput and memory capacity. Increas- of the under-utilized die space and the low storage require- ing computing density and per-server memory capacity al- ments of page-grain tags (5 MB for tags in a 4 GB soCache lows datacenter operators to deploy fewer servers for the module). The page-based design with in-SRAM tags offers same throughput and dataset, thus lowering cost [17, 31]. fundamental storage and design complexity advantages over We quantify the memory bandwidth and capacity require- state-of-the-art DRAM cache proposals that suffer from high ments of a scale-out server for various manufacturing tech- tag and metadata overheads that mandate in-DRAM storage. nologies (denoted by their year) in Table 1. The modeled organization is similar to emerging manycore servers, such Summarizing, our contributions are as follows: as Cavium’s 48-core processor fabricated in the 28nm tech- • Scale-out workloads exhibit skewed memory access nology [32]. Our configuration maximizes throughput by distributions, but high-capacity caches are needed to integrating maximum number of cores for a given die area capture the hot working sets. Due to disparity be- and power budget of 250-280 mm2 and 95-115 Watt. Our tween cache capacity and hot working set sizes, on-die models for future technologies are derived from ITRS [41]. stacked DRAM caches exhibit low temporal reuse — We estimate required memory capacity by examining var- e.g., a cache of 1 GB filters 45% of accesses in systems ious datacenter deployments. System vendors anticipate the with 100s of GB. The combination of poor cache per- need for cost-effective high server memory capacity, and formance and technological complexity makes on-die sell extended memory technology, which relies on multiple stacked DRAM caches unattractive for servers. buffer-on-board chips. Data analytic engines are provisioned with 2-8 GB per core [18], web search engines deploy 64 • High-capacity caches — enabled by the emergence GB for 16 cores [80] while web and streaming servers re- of high-bandwidth memory modules — are funda- quire 1-2 GB per core [20, 27]. Thus, we assume that 4 GB mentally different than on-die stacked DRAM caches. of per-core memory can be deployed cost-effectively. High-capacity caches access memory at coarse granu- We measure processor’s off-chip bandwidth demands by larity, making the case for simple page-based organi- scaling the per-core bandwidth consumption with the num- zations. Such organizations avoid the high metadata ber of cores available on the chip. We measure per-core storage requirements and excessive design complexity bandwidth by simulating a 16-core server (Section 5.2 de- of state-of-the-art DRAM caches [44, 45, 46, 60, 81]. tails the configuration). Our study shows that their per-core • soCache – a practical, scalable, and effective cache ar- bandwidth ranges from 0.4 GB/s to 1.2 GB/s corroborating chitecture for scale-out servers.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    15 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us