Proceedings of the 52nd Annual IEEE/ACM International Symposium on Microarchitecture 2019
DOI: 10.1145/3352460.3358300
|View full text |Cite
|
Sign up to set email alerts
|

Temporal Prefetching Without the Off-Chip Metadata

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
17
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 33 publications
(17 citation statements)
references
References 28 publications
0
17
0
Order By: Relevance
“…Thus, cold start mitigation strategies should prioritize unloading functions from memory when they are less likely to be invoked, rather than unloading them only when other functions need space. Hence, as demonstrated by other approaches on managing variable sized caches, Time-to-Live caches cannot be applied [4ś6, 52,69]. Also, traditional caching algorithms depends on the hit and miss ratios of all objects [13,17], whereas such a centralized control will not scale well for serverless functions.…”
Section: Related Workmentioning
confidence: 99%
“…Thus, cold start mitigation strategies should prioritize unloading functions from memory when they are less likely to be invoked, rather than unloading them only when other functions need space. Hence, as demonstrated by other approaches on managing variable sized caches, Time-to-Live caches cannot be applied [4ś6, 52,69]. Also, traditional caching algorithms depends on the hit and miss ratios of all objects [13,17], whereas such a centralized control will not scale well for serverless functions.…”
Section: Related Workmentioning
confidence: 99%
“…Temporal prefetchers usually demand hundreds of KBs of storage, which demands the storage of prefetch metadata in the off-chip memory. Some of the recent works on temporal prefetching are in the pursuit of improving the storage overhead without affecting the prefetch coverage [58], [59]. Berti, on the other hand, incurs a storage overhead of just 2.55 KB per core.…”
Section: Related Workmentioning
confidence: 99%
“…Accelerating irregular workloads using hardware prefetchers [37], [54]- [56], [73], [77], [95], [99] has been long studied that cover other types of data structures and memory access patterns containing linked lists, binary trees, hash joins in application domains such as geometric and scientific computations, high-performance computing, and databases. Furthermore, several temporal prefetchers [46], [93], [95], [96] and non-temporal prefetchers [13], [17], [52], [53], [64], [82], [86] are also investigated for these workloads. These approaches however, when applied in the graph processing context, can either prefetch for a subset of data structures or incur high complexity and cost for generality.…”
Section: Related Workmentioning
confidence: 99%