2012 SC Companion: High Performance Computing, Networking Storage and Analysis 2012
DOI: 10.1109/sc.companion.2012.43
|View full text |Cite
|
Sign up to set email alerts
|

Executing Optimized Irregular Applications Using Task Graphs within Existing Parallel Models

Abstract: Abstract-Many sparse or irregular scientific computations are memory bound and benefit from locality improving optimizations such as blocking or tiling. These optimizations result in asynchronous parallelism that can be represented by arbitrary task graphs. Unfortunately, most popular parallel programming models with the exception of Threading Building Blocks (TBB) do not directly execute arbitrary task graphs. In this paper, we compare the programming and execution of arbitrary task graphs qualitatively and q… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2013
2013
2023
2023

Publication Types

Select...
5

Relationship

1
4

Authors

Journals

citations
Cited by 6 publications
(2 citation statements)
references
References 20 publications
0
2
0
Order By: Relevance
“…Based on these patterns, a new schedule for execution is created and passed to an executor that runs the code according to the schedule generated by the inspector. Full sparse tiling (FST) [23], [24], [25] is an I/E optimization that improves temporal and spatial locality by placing the execution of loop iterations that access the same data, even across different original loops, together into a scheduling entity called a sparse tile. The tiles together perform the same computation iterations as the original code, just in an optimized order.…”
Section: A Full Sparse Tiling Overviewmentioning
confidence: 99%
See 1 more Smart Citation
“…Based on these patterns, a new schedule for execution is created and passed to an executor that runs the code according to the schedule generated by the inspector. Full sparse tiling (FST) [23], [24], [25] is an I/E optimization that improves temporal and spatial locality by placing the execution of loop iterations that access the same data, even across different original loops, together into a scheduling entity called a sparse tile. The tiles together perform the same computation iterations as the original code, just in an optimized order.…”
Section: A Full Sparse Tiling Overviewmentioning
confidence: 99%
“…For unstructured codes, there has been various inspector/executor strategies [49] that reschedule across loops to improve data locality while still providing parallelism [50], [24], [51], [25]. The term communication avoidance was coined by Demmel et al [51] to refer to such schedules, some of which have processors perform some amount of overlapped computation to avoid communication.…”
Section: Communication Avoidancementioning
confidence: 99%