Proceedings of the 1st Workshop on Architectures and Systems for Big Data 2011
DOI: 10.1145/2377978.2377981
|View full text |Cite
|
Sign up to set email alerts
|

Extending MPI to accelerators

Abstract: Current trends in computing and system architecture point towards a need for accelerators such as GPUs to have inherent communication capabilities. We review previous and current software libraries that provide pseudo-communication abilities through direct message passing. We show how these libraries are beneficial to the HPC community, but are not forward-thinking enough. We give motivation as to why MPI should be extended to support these accelerators, and provide a road map of achievable milestones to compl… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
8
0

Year Published

2012
2012
2021
2021

Publication Types

Select...
5
1
1

Relationship

3
4

Authors

Journals

citations
Cited by 15 publications
(8 citation statements)
references
References 5 publications
0
8
0
Order By: Relevance
“…Then, MVAPICH2-GPU internally uses different implementations depending on whether the memory buffer is in the device memory or the host memory. Stuart et al have discussed cl int clEnqueueReadBuffer( cl command queue cmd, /* command queue */ cl mem buf, /* memory buffer */ cl bool blocking, / blocking */ size t offset, /* offset */ size t size, /* buffer size */ void* hbuf, /* buffer pointer */ cl uint numevts, /* the number of events in the list */ cl event* wlist, /* event list */ cl evett* evtret ) /* event object of event object */ Algorithm 1 various design options of MPI extension to support accelerators [7]. Gelado et al proposed GMAC that provides a single memory space shared by a CPU and a GPU and hence allows MPI functions to access device memory data [8].…”
Section: Related Workmentioning
confidence: 99%
“…Then, MVAPICH2-GPU internally uses different implementations depending on whether the memory buffer is in the device memory or the host memory. Stuart et al have discussed cl int clEnqueueReadBuffer( cl command queue cmd, /* command queue */ cl mem buf, /* memory buffer */ cl bool blocking, / blocking */ size t offset, /* offset */ size t size, /* buffer size */ void* hbuf, /* buffer pointer */ cl uint numevts, /* the number of events in the list */ cl event* wlist, /* event list */ cl evett* evtret ) /* event object of event object */ Algorithm 1 various design options of MPI extension to support accelerators [7]. Gelado et al proposed GMAC that provides a single memory space shared by a CPU and a GPU and hence allows MPI functions to access device memory data [8].…”
Section: Related Workmentioning
confidence: 99%
“…Currently, only processes running on the CPU can perform MPI calls. Stuart et al [15] have suggested several mechanisms for extending the MPI standard to provide native support for accelerators. One significant proposal would allow GPU threads to obtain MPI ranks and participate directly in MPI communication [16].…”
Section: Related Workmentioning
confidence: 99%
“…Recently, Stuart et al proposed several potential directions for extending the MPI standard to provide native support of these accelerators [13]. One significant propsed extension is to allow accelerators to obtain MPI ranks and participate directly in MPI operations.…”
Section: Related Workmentioning
confidence: 99%