2004
DOI: 10.1109/tpds.2004.55
|View full text |Cite
|
Sign up to set email alerts
|

Finite state machine-based optimization of data parallel regular domain problems applied in low-level image processing

Abstract: A popular approach to providing nonexperts in parallel computing with an easy-to-use programming model is to design a software library consisting of a set of preparallelized routines, and hide the intricacies of parallelization behind the library's API. However, for regular domain problems (such as simple matrix manipulations or low-level image processing applications-in which all elements in a regular subset of a dense data field are accessed in turn) speedup obtained with many such library-based parallelizat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
11
0

Year Published

2004
2004
2011
2011

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 17 publications
(12 citation statements)
references
References 30 publications
(39 reference statements)
1
11
0
Order By: Relevance
“…Moreover, it was realized that it is not sufficient to consider parallelization of library operations in isolation. Therefore, the programming systems incorporate a run-time approach for communication minimization (called lazy parallelization) that automatically parallelizes a fully sequential program at runtime by inserting communication primitives and additional memory management operations whenever necessary [43].…”
Section: High-performance Distributed Multimedia Analysis With Jorusmentioning
confidence: 99%
“…Moreover, it was realized that it is not sufficient to consider parallelization of library operations in isolation. Therefore, the programming systems incorporate a run-time approach for communication minimization (called lazy parallelization) that automatically parallelizes a fully sequential program at runtime by inserting communication primitives and additional memory management operations whenever necessary [43].…”
Section: High-performance Distributed Multimedia Analysis With Jorusmentioning
confidence: 99%
“…Moreover, it was realized that it is not sufficient to consider parallelization of library operations in isolation. Therefore, the library was extended with a run-time approach for communication minimization (called lazy parallelization) that automatically parallelizes a fully sequential program at run-time by inserting communication primitives and additional memory management operations whenever necessary [10].…”
Section: Parallel-horusmentioning
confidence: 99%
“…This is explained by the fact that, in the case of horizontal partitioning, most of the data structures to be communicated are stored contiguously in memory. In contrast, vertical partitioning causes data structures to be stored non-contiguously, a problem which is discussed extensively in [16]. As a result, horizontal partitioning is our parallelization strategy of choice for each single iteration of the Householder bidiagonalization (see Figure 2).…”
Section: Parallelizing a Single Iterationmentioning
confidence: 99%