2010 ACM/IEEE International Conference for High Performance Computing, Networking, Storage and Analysis 2010
DOI: 10.1109/sc.2010.14
|View full text |Cite
|
Sign up to set email alerts
|

Combined Iterative and Model-driven Optimization in an Automatic Parallelization Framework

Abstract: Today's multi-core era places significant demands on an optimizing compiler, which must parallelize programs, exploit memory hierarchy, and leverage the ever-increasing SIMD capabilities of modern processors. Existing model-based heuristics for performance optimization used in compilers are limited in their ability to identify profitable parallelism/locality trade-offs and usually lead to sub-optimal performance.To address this problem, we distinguish optimizations for which effective model-based heuristics an… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
36
0

Year Published

2011
2011
2019
2019

Publication Types

Select...
3
3
1

Relationship

1
6

Authors

Journals

citations
Cited by 50 publications
(36 citation statements)
references
References 31 publications
0
36
0
Order By: Relevance
“…However, the space of possible optimizations is dramatically enlarged, imposing challenges on the selection algorithms. In contrast to previous iterative approaches requiring the evaluation of up to hundreds of possible transformations [30], [32], we develop here a scheme that requires at most 5 candidate choices to be evaluated.…”
Section: Optimization Spacementioning
confidence: 99%
See 2 more Smart Citations
“…However, the space of possible optimizations is dramatically enlarged, imposing challenges on the selection algorithms. In contrast to previous iterative approaches requiring the evaluation of up to hundreds of possible transformations [30], [32], we develop here a scheme that requires at most 5 candidate choices to be evaluated.…”
Section: Optimization Spacementioning
confidence: 99%
“…The best loop optimization sequence is often not only program-specific, but also depends on the target hardware. Pouchet et al illustrated this by showing the critical impact of tuning polyhedral optimizations for obtaining the best performance for a variety of numerical programs on different target processors [31], [30], [32].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Since MLP is an indicator of intra-warp memory-level parallelism, we need to consider the overlap factor of multiple warps. ITMLP can be calculated using Equations (14) and (15).…”
Section: Calculating the Memory Access Cost Tmemmentioning
confidence: 99%
“…Our work is also related to a rich body of work on optimizations and tuning of GPGPU applications [4,7,12,15,16,19]. Ryoo et al [16] introduced two metrics to prune optimization space by calculating the utilization and efficiency of GPU applications.…”
Section: Gpu Performance Modelingmentioning
confidence: 99%