2017
DOI: 10.1007/978-3-319-65578-9_4
|View full text |Cite
|
Sign up to set email alerts
|

Compiling and Optimizing OpenMP 4.X Programs to OpenCL and SPIR

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
7
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
2
2
2

Relationship

2
4

Authors

Journals

citations
Cited by 10 publications
(7 citation statements)
references
References 14 publications
0
7
0
Order By: Relevance
“…fork/join. Pereira et al [16] describe a framework that automatically converts program sections annotated with OpenMP 4.x directives into OpenCL kernels. This design goes in the direction of our work, but we perform a step further, considering the severe constraints of the ULP MCUs, requiring specific optimizations.…”
Section: Related Workmentioning
confidence: 99%
“…fork/join. Pereira et al [16] describe a framework that automatically converts program sections annotated with OpenMP 4.x directives into OpenCL kernels. This design goes in the direction of our work, but we perform a step further, considering the severe constraints of the ULP MCUs, requiring specific optimizations.…”
Section: Related Workmentioning
confidence: 99%
“…Most of the rest of the literature that strictly follows the OpenMP accelerator model studied GPU offloading and demonstrate good results [23,24]. Nevertheless, some of them look for more untraditional targets like the Intel Xeon Phi platform or FPGAs [65].…”
Section: Related Workmentioning
confidence: 99%
“…The map clause details the mapping of the data between the host and the target device: inputs (A and B) are mapped to the target, and the output (C) is mapped from the target. While typical target devices are DSP cores, GPUs, Xeon Phi accelerators, and so on [23,24], this article introduces the cloud as yet another target device available from the local computer, giving the programmer the ability to quickly expand the computational power of its own computer to a large-scale cloud cluster. Using OmpCloud, the programmer can leverage on his/her basic OpenMP knowledge.…”
Section: Introductionmentioning
confidence: 99%
“…While those works demonstrate good performance, Xeon Phi are quite an exotic platforms not really available to the majority of users. Most of the rest of the literature which strictly follows the OpenMP accelerator model studied GPU offloading and demonstrate good results for heavily parallel applications [8], [9]. Nevertheless, some of them look for more untraditional targets like the Intel Xeon Phi platform or FPGAs [24].…”
Section: Related Workmentioning
confidence: 99%
“…While the typical devices used in OpenMP 4.X are DSP cores, GPUs, Xeon Phi accelerators, etc. [8], [9], we introduced the cloud as a novel target device available on the computer. This was done within a programming framework we call OmpCloud [3] that extends the OpenMP accelerator model to allow transparent cloud offloading and cluster programming.…”
Section: Introductionmentioning
confidence: 99%