Proceedings of the 2004 ACM Symposium on Applied Computing 2004
DOI: 10.1145/967900.968184
|View full text |Cite
|
Sign up to set email alerts
|

Automatic parallel code generation for tiled nested loops

Abstract: This paper presents an overview of our work, concerning a complete end-to-end framework for automatically generating message passing parallel code for tiled nested for-loops. It considers general parallelepiped tiling transformations and general convex iteration spaces. We address all problems regarding both the generation of sequential tiled code and its parallelization. We have implemented our techniques in a tool which automatically generates MPI parallel code and conducted several series of experiments, co… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2006
2006
2016
2016

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(6 citation statements)
references
References 27 publications
(43 reference statements)
0
6
0
Order By: Relevance
“…In the digital signal processing domain, Bondalapati [7] try to parallelize nested loops. In [24], the authors try to automatically parallelize the tiled loop nests using a messagepassing interface (MPI [22]). Loop-level parallelism for coarsegrained reconfigurable architectures is introduced in [43], while Hogstedt et al [27] investigate the parallel execution time of tiled loop nests.…”
Section: Discussion Of Related Workmentioning
confidence: 99%
“…In the digital signal processing domain, Bondalapati [7] try to parallelize nested loops. In [24], the authors try to automatically parallelize the tiled loop nests using a messagepassing interface (MPI [22]). Loop-level parallelism for coarsegrained reconfigurable architectures is introduced in [43], while Hogstedt et al [27] investigate the parallel execution time of tiled loop nests.…”
Section: Discussion Of Related Workmentioning
confidence: 99%
“…To parallelize such loops, they exploit the distributed memory available in the reconfigurable architecture by implementing a data context switching technique. Goumas et al [36] propose a framework to automatically generate parallel code for tiled nested loops. They have implemented several loop transformations within the proposed approach using MPI [22], the message-passing parallel interface.…”
Section: Relevant Prior Workmentioning
confidence: 99%
“…This introduces rescaling to q processors. Next we apply rules (9), (14), and (16)- (18) to formally parallelize:…”
Section: Rescaling Ffts Using Spiralmentioning
confidence: 99%
“…A compiler framework for generating MPI code for arbitrarily tiled for-loop nests by performing various loop transformations to gain inherent coarse-grained parallelism is presented in [14]. [18] describes the generation of collective communication MPI code by automatically searching for the best algorithm on a given system.…”
Section: Introductionmentioning
confidence: 99%