2012
DOI: 10.1007/978-3-642-32820-6_85
|View full text |Cite
|
Sign up to set email alerts
|

OpenACC — First Experiences with Real-World Applications

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
120
0
2

Year Published

2014
2014
2015
2015

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 216 publications
(122 citation statements)
references
References 8 publications
0
120
0
2
Order By: Relevance
“…Previous studies, like [8,24,13,17,32,14,22,10] also evaluate directive-based compilers that generate code for accelerators. The main difference is that this work covers more programs and includes a study of transformations.…”
Section: Related Workmentioning
confidence: 99%
“…Previous studies, like [8,24,13,17,32,14,22,10] also evaluate directive-based compilers that generate code for accelerators. The main difference is that this work covers more programs and includes a study of transformations.…”
Section: Related Workmentioning
confidence: 99%
“…5. Compute W rong A in parallel (line [14][15][16], which indicates the incorrect elements of array A. An element is incorrect only when it is written by at least one misspeculated iteration.…”
Section: Irregular Memory Accessesmentioning
confidence: 99%
“…Several programming models have been proposed for GPU computing including OpenCL [15], CUDA [12], OpenACC [16], PGI Accelerator [17], OmpSs [2] which is based upon OpenMP standard [4], and Par4All [1]. None of these programming models support speculative parallelization for GPU computing.…”
Section: Related Workmentioning
confidence: 99%
“…Also, recent tools such as Kernelgen [34], Polly [24], hiCUDA [25], and the GPSME toolkit [57] are gaining in popularity and successful use-cases. Successful examples of semi-automatically parallelized programs range from medicine [56] to physics simulations [33]. Some of the tools only make use of the raw computational power of accelerators, while other tools offer implicit or explicit support for optimizing the usage of the GPU's complex memory hierarchy and hence minimizing communication [6].…”
Section: Automatic Parallelizationmentioning
confidence: 99%