Proceedings of the 53rd Annual Design Automation Conference 2016
DOI: 10.1145/2897937.2898036
|View full text |Cite
|
Sign up to set email alerts
|

Latency sensitivity-based cache partitioning for heterogeneous multi-core architecture

Abstract: Shared last-level cache (LLC) management is a critical design issue for heterogeneous multi-cores. In this paper, we observe two major challenges: the contribution of LLC latency to overall performance varies among applications/cores and also across time; overlooking the off-chip latency factor often leads to adverse effects on overall performance. Hence, we propose a Latency Sensitivity-based Cache Partitioning (LSP) framework, including a lightweight runtime mechanism to quantify the latency-sensitivity and … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
7
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 19 publications
(7 citation statements)
references
References 19 publications
0
7
0
Order By: Relevance
“…Therefore, the LLC design is a very important issue in multi-core systems. Wang et al [218] have proposed a latency sensitivity-based cache partitioning (LSP) framework. The LSP framework, evaluates a latencysensitivity metric at runtime to adapt the cache partitioning.…”
Section: ) Multi-core Optimizationmentioning
confidence: 99%
“…Therefore, the LLC design is a very important issue in multi-core systems. Wang et al [218] have proposed a latency sensitivity-based cache partitioning (LSP) framework. The LSP framework, evaluates a latencysensitivity metric at runtime to adapt the cache partitioning.…”
Section: ) Multi-core Optimizationmentioning
confidence: 99%
“…Further, the single-tier virtual queuing memory controller [5] was proposed to overcome the limitation of two-tier schedulers in QoS-aware scheduling. Besides memory scheduling, QoSaware cache management [7] and on-chip network design [6] have also been well-explored in recent years. Nonetheless, these work cannot guarantee end-to-end QoS because they only deal with certain parts of the memory system.…”
Section: Related Workmentioning
confidence: 99%
“…Moreover, as latency-sensitive cores such as the DSP share memory with other cores, they can be easily overwhelmed by real-time cores consuming high bandwidth. QoS-aware management for specific types of memory resources has been well-studied by previous work [3,4,5,6,7]. In [3], a QoS-aware scheduling policy was proposed for CPU-GPU systems.…”
Section: Introductionmentioning
confidence: 99%
“…QoS-aware management for specific types of memory resources has been well studied by previous work [1,4,9,22,25]. In Reference [9], a QoS-aware scheduling policy was proposed for CPU-GPU systems.…”
Section: Introductionmentioning
confidence: 99%
“…With ineffective memory scheduling, a real-time core (e.g., the display) may not achieve the target real-time performance due to inadequate memory bandwidth. Moreover, as latency-sensitive cores such as the DSP share memory with other cores, they can be easily overwhelmed by real-time cores consuming high bandwidth.QoS-aware management for specific types of memory resources has been well studied by previous work [1,4,9,22,25]. In Reference [9], a QoS-aware scheduling policy was proposed for CPU-GPU systems.…”
mentioning
confidence: 99%