2015
DOI: 10.1142/s0129626415500048
|View full text |Cite
|
Sign up to set email alerts
|

A Framework for Efficient Execution of Data Parallel Irregular Applications on Heterogeneous Systems

Abstract: Exploiting the computing power of the diversity of resources available on heterogeneous systems is mandatory but a very challenging task. The diversity of architectures, execution models and programming tools, together with disjoint address spaces and different computing capabilities, raise a number of challenges that severely impact on application performance and programming productivity. This problem is further compounded in the presence of data parallel irregular applications. This paper presents a framewo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 13 publications
(24 reference statements)
0
3
0
Order By: Relevance
“…Barbosa et al [5,18] proposed allowing the developer to define the partition method for the data and rely on a performance model and scheduler to dice the tasks to be decomposed into smaller ones when needed. This mechanism can be leveraged to enable the interweaving of simulation and visualization tasks, primarily because they operate in different time intervals, t and t − 1, respectively.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Barbosa et al [5,18] proposed allowing the developer to define the partition method for the data and rely on a performance model and scheduler to dice the tasks to be decomposed into smaller ones when needed. This mechanism can be leveraged to enable the interweaving of simulation and visualization tasks, primarily because they operate in different time intervals, t and t − 1, respectively.…”
Section: Related Workmentioning
confidence: 99%
“…The solution gives rise to a granularity of τ v , i.e., if the workload from τ v (W (τ v )) is too large, it will delay the execution of subsequently scheduled τ s , on the other hand, if W (τ v ) is too small it will increase the scheduling overhead. To address this problem, we follow the approach proposed by Barbosa et al [5,18] referred to as dicing. The dicing strategy enables the developer to define a generic workload applied to a partition created by the scheduler using a developer-defined partitioning method called dice.…”
Section: Our Approachmentioning
confidence: 99%
“…Furthermore, different architectures usually exhibit different execution and programming models and are deployed with different programming languages and development tools, severely impacting on both code and performance portability. Additionally, the application's workload has to be distributed and balanced among the multiple devices, and, within each device, among its multiple computing units; this leads to multilevel scheduling, which must be effectively handled in order to achieve acceptable performance levels [1]. To efficiently use the available resources in these and future systems, algorithms and software packages have to be revisited and re-evaluated to assess their adequateness to these environments.…”
Section: Introductionmentioning
confidence: 99%