2016 IEEE International Symposium on High Performance Computer Architecture (HPCA) 2016
DOI: 10.1109/hpca.2016.7446065
|View full text |Cite
|
Sign up to set email alerts
|

Minimal disturbance placement and promotion

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
17
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
4
2

Relationship

2
4

Authors

Journals

citations
Cited by 12 publications
(17 citation statements)
references
References 23 publications
0
17
0
Order By: Relevance
“…This is in contrast with prior predictors [2], [4]- [6], which need to access a dedicated predictor table upon every single LLC access. Because modern multicore processors feature distributed last-level caches, accesses to dedicated prediction tables introduce detrimental latency and energy overheads in traversing the on-chip interconnect to query such structures.…”
Section: Introductionmentioning
confidence: 94%
See 1 more Smart Citation
“…This is in contrast with prior predictors [2], [4]- [6], which need to access a dedicated predictor table upon every single LLC access. Because modern multicore processors feature distributed last-level caches, accesses to dedicated prediction tables introduce detrimental latency and energy overheads in traversing the on-chip interconnect to query such structures.…”
Section: Introductionmentioning
confidence: 94%
“…We evaluate the performance of SPEC CPU 2006 benchmarks using a modified version of CMP$im [15] provided with the JILP Cache Replacement Championship [16] and used in prior research in dead block prediction [2], [4], [5], [14]. Table III summarizes the features of the simulated processor.…”
Section: Spec Cpu 2006mentioning
confidence: 99%
“…2 History-based predictive schemes such as the state-ofthe-art Hawkeye [26] and many others [5,10,13,28,29,49,53] learn past reuse behavior of cache blocks by employing sophisticated storage-intensive prediction mechanisms. A large body of recent works focus on history-based schemes as they generally provide higher performance than the lightweight schemes for a wide range of applications.…”
Section: F Prior Hardware Schemesmentioning
confidence: 99%
“…These hardware schemes aim to perform two tasks: (1) identify cache blocks that are likely to exhibit high reuse, and (2) protect high reuse cache blocks from cache thrashing. To accomplish the first task, these schemes deploy either probabilistic or prediction-based hardware mechanisms [5,10,13,26,28,29,41,49,51,52,53,57,58,59,60]. However, our work finds that graph-dependent irregular access patterns prevent these schemes from correctly learning which cache blocks to preserve, rendering them deficient for the broad domain of graph analytics.…”
Section: Introductionmentioning
confidence: 96%
“…With so much of the available on-die resources invested in the cache hierarchy, an efficient, high performance design requires intelligent cache management techniques. While many cache management and speculation techniques such as alternate replacement policies [6,14,15,18,27], deadblock/hit prediction [17,20,28,33,36], and prefetching techniques [2,10,16,19,19,23,26,31,32] have been extensively explored, many of these are piecemeal, one-off solutions that often interact poorly when implemented together and typically only address one level of the memory-system hierarchy. There has been little work exploring the interactions between these policies across multiple levels of the memory hierarchy and examining the information needed across boundaries in the system from software to the core, to the last level cache.…”
Section: Introductionmentioning
confidence: 99%