2009 IEEE International Symposium on Performance Analysis of Systems and Software 2009
DOI: 10.1109/ispass.2009.4919641
|View full text |Cite
|
Sign up to set email alerts
|

Machine learning based online performance prediction for runtime parallelization and task scheduling

Abstract: Abstract-With the emerging many-core paradigm, parallel programming must extend beyond its traditional realm of scientific applications. Converting existing sequential applications as well as developing next-generation software requires assistance from hardware, compilers and runtime systems to exploit parallelism transparently within applications. These systems must decompose applications into tasks that can be executed in parallel and then schedule those tasks to minimize load imbalance. However, many system… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
24
0

Year Published

2011
2011
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 44 publications
(24 citation statements)
references
References 38 publications
0
24
0
Order By: Relevance
“…Li, et al [26], similarly to our work, use ANNs as models for online performance prediction, that they apply to task partitioning and scheduling for HPC clusters. Their work does not consider NUMA architectures, core and data layout effects, or power metrics.…”
Section: Related Workmentioning
confidence: 90%
See 1 more Smart Citation
“…Li, et al [26], similarly to our work, use ANNs as models for online performance prediction, that they apply to task partitioning and scheduling for HPC clusters. Their work does not consider NUMA architectures, core and data layout effects, or power metrics.…”
Section: Related Workmentioning
confidence: 90%
“…Li, et al, [26] and Ipek, et al, [27] use artificial neural networks (ANNs) as blackbox models for microarchitectural design space exploration. Li, et al [26], similarly to our work, use ANNs as models for online performance prediction, that they apply to task partitioning and scheduling for HPC clusters.…”
Section: Related Workmentioning
confidence: 99%
“…They measured assignments, branches, and loops at run-time using dynamic analysis of the program. In [22], non-deterministic features were measured. The variables ar, ot were represented using performance counters: number of CPU cycles, number of cache misses, cache accesses for the last cache level, and number of level one cache hits.…”
Section: Data Generation and Model Selectionmentioning
confidence: 99%
“…In supervised learning, on the other hand, the algorithm is provided Best tuning params. 13,15,20,21,28,29 7,29 24,30,31 Work distribution 25,34 33,34 Power 23,27 Output size 17 between the input and output variables of interest. These data are then given to a machine learning algorithm that builds a model.…”
Section: Machine Learning For Performance Modelingmentioning
confidence: 99%