Proceedings of the International Symposium on Memory Systems 2018
DOI: 10.1145/3240302.3240310
|View full text |Cite
|
Sign up to set email alerts
|

Design guidelines for high-performance SCM hierarchies

Abstract: With emerging storage-class memory (SCM) nearing commercialization, there is evidence that it will deliver the much-anticipated high density and access latencies within only a few factors of DRAM. Nevertheless, the latency-sensitive nature of memory-resident services makes seamless integration of SCM in servers questionable. In this paper, we ask the question of how best to introduce SCM for such servers to improve overall performance/cost over existing DRAM-only architectures. We first show that even with the… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
11
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 12 publications
(11 citation statements)
references
References 65 publications
0
11
0
Order By: Relevance
“…However, such a scheme suffers from long latencies and low throughput due to high overhead from the software stack. Prior work such as [18,20,45] has tackled this high overhead by enabling cache line access to the SSD [18,20], using host DRAM as a cache [45], merging multiple translation layers into one layer, and promoting pages from SSD when locality is detected [18]. Due to its benefits, Flash memory is abundantly available, extensively used, and continues to be optimized to provide better performance [13,22,35].…”
Section: Flash As Memorymentioning
confidence: 99%
“…However, such a scheme suffers from long latencies and low throughput due to high overhead from the software stack. Prior work such as [18,20,45] has tackled this high overhead by enabling cache line access to the SSD [18,20], using host DRAM as a cache [45], merging multiple translation layers into one layer, and promoting pages from SSD when locality is detected [18]. Due to its benefits, Flash memory is abundantly available, extensively used, and continues to be optimized to provide better performance [13,22,35].…”
Section: Flash As Memorymentioning
confidence: 99%
“…First, many modern online services have msscale end-to-end tail-latency constraints [18], [23], which allows them to absorb few µs-scale flash accesses [14], [18], [22], [41], [42], [44]. Second, object popularity and request distributions for datacenter workloads are inherently skewed [64], [73], [75], [76], thus allowing hosting the hot fraction of the dataset in DRAM that serves most requests and filters the bandwidth required from the backing flash. The above observations should permit the design of a cost-effective two-tier hierarchy where a capacity-constrained DRAM caches the hot fraction of the dataset stored in a capacity-scaled flash layer.…”
Section: Introductionmentioning
confidence: 99%
“…This approach allows for the memory to be scaled out independently of the processing resources, because scaling up the memory within a node is becoming increasingly prohibitive [108]. This "share something" approach for remote memory targets some of the challenges that are very relevant in today's data centers, some of which that are:…”
Section: Memory Management Across Virtualized Nodesmentioning
confidence: 99%
“…When memory is disaggregated and becomes a pooled resource, it is possible to use different types of memory technologies besides the traditional DRAM, such as Storage-Class Memories (SCMs) [108], a term that refers generally to Non-Volatile Memories (NVMs) [77]. NVMs provide a good option for pooled disaggregated memory because it is able to provide higher density and higher capacity than traditional DRAMs, but at a much lower cost [28].…”
Section: The Memory Management Problem With Virtualization Andmentioning
confidence: 99%
See 1 more Smart Citation