2014
DOI: 10.1007/978-3-642-54807-9_9
|View full text |Cite
|
Sign up to set email alerts
|

Exploitation of GPUs for the Parallelisation of Probably Parallel Legacy Code

Abstract: General purpose Gpus provide massive compute power, but are notoriously difficult to program. In this paper we present a complete compilation strategy to exploit Gpus for the parallelisation of sequential legacy code. Using hybrid data dependence analysis combining static and dynamic information, our compiler automatically detects suitable parallelism and generates parallel OpenCl code from sequential programs. We exploit the fact that dependence profiling provides us with parallel loop candidates that are hig… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2014
2014
2023
2023

Publication Types

Select...
5
3

Relationship

6
2

Authors

Journals

citations
Cited by 16 publications
(11 citation statements)
references
References 28 publications
0
10
0
Order By: Relevance
“…In all preceding approaches, the code always gets executed on the GPU. Prior work on automatic generation of parallel GPU code from sequential programs also includes Par4ALL [Amini et al 2012], PPCG [Verdoolaege et al 2013], and that of Wang et al [2014a]. Unlike our approach, they do not consider the problem of selecting the most suitable device from the host CPU and the GPU to run the code.…”
Section: Related Workmentioning
confidence: 91%
“…In all preceding approaches, the code always gets executed on the GPU. Prior work on automatic generation of parallel GPU code from sequential programs also includes Par4ALL [Amini et al 2012], PPCG [Verdoolaege et al 2013], and that of Wang et al [2014a]. Unlike our approach, they do not consider the problem of selecting the most suitable device from the host CPU and the GPU to run the code.…”
Section: Related Workmentioning
confidence: 91%
“…The Open-MPC compiler [69] translates OpenMP to CUDA programs. Wang et al [20], [24], [70] translates OpenMP to OpenCL programs and use machine learning to select the most suitable device from the host CPU and the GPU to run the code. Rawat et al presents an automatic approach to generate GPU code from a domain-specific language (DSL) for stencil programs [71].…”
Section: Domain-specific Optimizationsmentioning
confidence: 99%
“…Machine learning has been employed for various optimization tasks [40], including code optimization [7,12,29,30,37,39,41,42,43,44,45,46,51], task scheduling [9,10,11,33,34], model selection [38], etc.…”
Section: Related Workmentioning
confidence: 99%