2012
DOI: 10.1007/978-3-642-33518-1_40
|View full text |Cite
|
Sign up to set email alerts
|

StarPU-MPI: Task Programming over Clusters of Machines Enhanced with Accelerators

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
56
0
1

Year Published

2013
2013
2018
2018

Publication Types

Select...
7
2
1

Relationship

2
8

Authors

Journals

citations
Cited by 47 publications
(61 citation statements)
references
References 3 publications
(2 reference statements)
0
56
0
1
Order By: Relevance
“…However, there has also been some efforts to apply task-based approaches to clusters of computers. Cilk-NOW is a variant of Cilk for networks of workstations [8], StarPU-MPI is an extension of StarPU for clusters of acceleratorenhanced machines [9], and OmpSs has been implemented for clusters of GPUs as well [5]. The DAGuE framework [10] is an example of a task-based approach that achieves high performance in dense linear algebra operations.…”
Section: Parallel Programming Modelsmentioning
confidence: 99%
“…However, there has also been some efforts to apply task-based approaches to clusters of computers. Cilk-NOW is a variant of Cilk for networks of workstations [8], StarPU-MPI is an extension of StarPU for clusters of acceleratorenhanced machines [9], and OmpSs has been implemented for clusters of GPUs as well [5]. The DAGuE framework [10] is an example of a task-based approach that achieves high performance in dense linear algebra operations.…”
Section: Parallel Programming Modelsmentioning
confidence: 99%
“…OMPSs may schedule program on a distributed memory architecture in the same fashion as our old Kaapi implementation [15]. StarPU [2] allows to mix dependent tasks with MPI commu- nication primitives. X-KAAPI provides a binary compatible libGOMP library called libKOMP [5].…”
Section: Overall Gains For Epxmentioning
confidence: 99%
“…Prior work on heterogeneous CPU/GPU systems has focused on new programming models and API extensions for supporting multiple heterogeneous devices [23,3,12], automating the mapping processor [5,15,16], enabling CPU and GPU sharing [20]. Different mapping heuristics have been designed and applied in these work.…”
Section: Introductionmentioning
confidence: 99%