2013 IEEE International Symposium on Parallel &Amp; Distributed Processing, Workshops and PHD Forum 2013
DOI: 10.1109/ipdpsw.2013.263
|View full text |Cite
|
Sign up to set email alerts
|

Exploring Programming Multi-GPUs Using OpenMP and OpenACC-Based Hybrid Model

Abstract: Heterogeneous computing come with tremendous potential and is a leading candidate for scientific applications that are becoming more and more complex. Accelerators such as GPUs whose computing momentum is growing faster than ever offer application performance when compute intensive portions of an application are offloaded to them. It is quite evident that future computing architectures are moving towards hybrid systems consisting of multi-GPUs and multicore CPUs. A variety of high-level languages and software … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
20
0
1

Year Published

2014
2014
2023
2023

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 20 publications
(21 citation statements)
references
References 12 publications
0
20
0
1
Order By: Relevance
“…In order to add parallel capacity for systems based on multi-core CPU, there are different platforms and libraries available-for example, Intel (Folsom, CA, USA) Thread Building Blocks (TBB) [9,10], Pthreads [10], Message Passing Interface (MPI) [11], OpenMP [11,12], Open Computing Language (OpenCL) [13] and Open Accelerators (OpenACC) [14]. All of them can be used for programming multi-core CPUs, but MPI is more appropriate for distributed shared memory (DSM) systems, such as computational clusters [11].…”
Section: Parallel Processingmentioning
confidence: 99%
See 1 more Smart Citation
“…In order to add parallel capacity for systems based on multi-core CPU, there are different platforms and libraries available-for example, Intel (Folsom, CA, USA) Thread Building Blocks (TBB) [9,10], Pthreads [10], Message Passing Interface (MPI) [11], OpenMP [11,12], Open Computing Language (OpenCL) [13] and Open Accelerators (OpenACC) [14]. All of them can be used for programming multi-core CPUs, but MPI is more appropriate for distributed shared memory (DSM) systems, such as computational clusters [11].…”
Section: Parallel Processingmentioning
confidence: 99%
“…Although presenting the advantage of a unified view of the system memory space, OpenMP does not have a better performance than MPI when used on heterogeneous architectures with non-uniform memory access (NUMA) [11,12]. Since version 4.0, OpenMP has supported programming for GPU [14].…”
Section: Parallel Processingmentioning
confidence: 99%
“…Although these high-level programming models lack such [20] used OpenMP + OpenACC to utilize multiple GPUs attached in a multi-core system (a single node of a cluster) and explained how to do the inter-GPU communication with OpenMP synchronization and OpenACC data synchronization primitives. Hart et al [8] used Co-Array Fortran (CAF) + OpenACC and Levesque et al [10] used MPI + OpenACC to utilize multi-GPU in GPU cluster.…”
Section: Related Workmentioning
confidence: 99%
“…The exploitation of heterogeneous parallelism across the devices existing in a node by combining OpenMP and OpenACC, or ad-hoc directives has been explored in [26] and in [5], respectively. The result and the differences with respect to HPL are similar to those of OmpSs, with the addition that these solutions do not automatically schedule and synchronize the tasks based on their data dependencies.…”
Section: Related Workmentioning
confidence: 99%