2019
DOI: 10.1145/3323055
|View full text |Cite
|
Sign up to set email alerts
|

Application and Thermal-reliability-aware Reinforcement Learning Based Multi-core Power Management

Abstract: Power management through dynamic voltage and frequency scaling (DVFS) is one of the most widely adopted techniques. However, it impacts application reliability (due to soft errors, circuit aging, and deadline misses). However, increased power density impacts the thermal reliability of the chip, sometimes leading to permanent failure. To balance both application- and thermal-reliability along with achieving power savings and maintaining performance, we propose application- and thermal-reliability-aware reinforc… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 21 publications
(6 citation statements)
references
References 26 publications
0
6
0
Order By: Relevance
“…Although the RL technique has been reported to be lightweight and highly suitable for the systems, compared to other types of learning techniques [28], the main issues are its convergence and timing overhead. Accordingly, similar to other studies [31], we have reduced the feasible actions to reduce the complexity and convergence issues. In the following, we investigate the timing and memory overheads of the employed learning technique.…”
Section: Investigating the Timing And Memory Overheads Of ML Techniquementioning
confidence: 88%
See 2 more Smart Citations
“…Although the RL technique has been reported to be lightweight and highly suitable for the systems, compared to other types of learning techniques [28], the main issues are its convergence and timing overhead. Accordingly, similar to other studies [31], we have reduced the feasible actions to reduce the complexity and convergence issues. In the following, we investigate the timing and memory overheads of the employed learning technique.…”
Section: Investigating the Timing And Memory Overheads Of ML Techniquementioning
confidence: 88%
“…The Q-learning/SARSA technique, which is recently been used in many emerging applications, such as robotics, and Unmanned Aerial Vehicles (UAV) [29,30], uses the RL technique to perform the runtime management/optimization of the system properties in single or multi-core processors. The general Q-learning/SARSA technique consists of the three main components [31,32], including: (1) a discrete set of states S = {s 1 , s 2 , ..., s l }, (2) a discrete set of actions A = {a 1 , a 2 , ..., a k }, and (3) reward function R. The states and actions determine the rows and columns of the Q-table of the learning-based algorithm, respectively (shown in Figure 2). The algorithm collects the current state s t , and determines the next action a t (a t ∈ A).…”
Section: Learning-based System Properties Optimizationmentioning
confidence: 99%
See 1 more Smart Citation
“…Several works have employed RL for power/thermal optimization [31]. The works in [18], [20], [21] use RL for power management via DVFS. However, they neither consider temperature nor QoS.…”
Section: Related Workmentioning
confidence: 99%
“…As a result, most DVFS techniques [32,36,40,42,48] relying on a predefined temperature prediction model would not work properly for mobile devices. Furthermore, recent supervised learning-based approaches [8,11,45,53] only give the adaptation ability to previously trained environments. Thus, their thermal management in mobile settings has no performance guarantee.…”
Section: Introductionmentioning
confidence: 99%