2009 IEEE International Conference on Cluster Computing and Workshops 2009
DOI: 10.1109/clustr.2009.5289129
|View full text |Cite
|
Sign up to set email alerts
|

Message passing for GPGPU clusters: CudaMPI

Abstract: Abstract-We present and analyze two new communication libraries, cudaMPI and glMPI, that provide an MPI-like message passing interface to communicate data stored on the graphics cards of a distributed-memory parallel computer. These libraries can help applications that perform general purpose computations on these networked GPU clusters. We explore how to efficiently support both point-to-point and collective communication for either contiguous or noncontiguous data on modern graphics cards. Our software desig… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
28
0

Year Published

2011
2011
2018
2018

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 43 publications
(28 citation statements)
references
References 9 publications
0
28
0
Order By: Relevance
“…DisMaRC, a distributed GPGPU based MapReduce framework is presented in [12]. In another work by Lawlor [13], the author analyses two new communication libraries, cudaMPI and glMPI, that provide an MPI-like message passing interface to communicate data stored on the graphics cards of a distributedmemory parallel computer. Then there are numerous examples of single applications that are ported on cluster of GPUs [14], [15], [16], [17].…”
Section: Related Workmentioning
confidence: 99%
“…DisMaRC, a distributed GPGPU based MapReduce framework is presented in [12]. In another work by Lawlor [13], the author analyses two new communication libraries, cudaMPI and glMPI, that provide an MPI-like message passing interface to communicate data stored on the graphics cards of a distributedmemory parallel computer. Then there are numerous examples of single applications that are ported on cluster of GPUs [14], [15], [16], [17].…”
Section: Related Workmentioning
confidence: 99%
“…The cudaMPI library studies providing wrapper API functions by mixing CUDA and MPI data movement [21]. Similarly to MPI-ACC, Wang et al propose to add CUDA [2] support to MVAPICH2 [22] and optimize the internode communication for InfiniBand networks [28].…”
Section: Related Workmentioning
confidence: 99%
“…cudaMPI [3] provided an MPI-like interface that handled the inherent memory copies from the GPU to the CPU before the CPU executed an MPI call. However, thanks to 0-copy memory 1 introduced in CUDA 3 [4], cudaMPI is redundant.…”
Section: Previous Workmentioning
confidence: 99%