2010
DOI: 10.1007/s11227-010-0418-y
|View full text |Cite
|
Sign up to set email alerts
|

Designing parallel loop self-scheduling schemes using the hybrid MPI and OpenMP programming model for multi-core grid systems

Abstract: Loop scheduling on parallel and distributed systems has been thoroughly investigated in the past. However, none of these studies considered the multi-core architecture feature for emerging grid systems. Although there have been many studies proposed to employ the hybrid MPI and OpenMP programming model to exploit different levels of parallelism for a distributed system with multi-core computers, none of them were aimed at parallel loop self-scheduling. Therefore, this paper investigates how to employ the hybri… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2012
2012
2019
2019

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 12 publications
(12 citation statements)
references
References 22 publications
0
12
0
Order By: Relevance
“…SMV is an application kernel from Sparse Linear Algebra that performs a multiplication between a sparse matrix and a dense vector. Besides finding applications in several scientific and engineering domains, sparse matrix‐vector multiplication is a frequently studied application kernel within the context of loop scheduling . We extracted the SMV kernel from the Conjugate Gradient application from the NAS Parallel Benchmarks (NPB) .…”
Section: Evaluation Methodologymentioning
confidence: 99%
See 1 more Smart Citation
“…SMV is an application kernel from Sparse Linear Algebra that performs a multiplication between a sparse matrix and a dense vector. Besides finding applications in several scientific and engineering domains, sparse matrix‐vector multiplication is a frequently studied application kernel within the context of loop scheduling . We extracted the SMV kernel from the Conjugate Gradient application from the NAS Parallel Benchmarks (NPB) .…”
Section: Evaluation Methodologymentioning
confidence: 99%
“…Besides finding applications in several scientific and engineering domains, 24 sparse matrix-vector multiplication is a frequently studied application kernel within the context of loop scheduling. 25,26 We extracted the SMV kernel from the Conjugate Gradient application from the NAS Parallel Benchmarks (NPB). 27 In the SMV kernel, the sparse matrix is stored in compressed row format so that memory can be saved and data affinity exploited.…”
Section: Application Kernelsmentioning
confidence: 99%
“…A common approach is to use MPI [1] for inter-node communication and OpenMP [4] for programming the shared-memory systems [6]. In the context of DLS techniques, the hierarchical loop scheduling (HLS) [14] was one of the earliest efforts to use MPI+OpenMP programming model. In HLS, a free worker (MPI process) requests a chunk from the master rank which calculates and assigns the chunk based on a certain performance function [30].…”
Section: Background and Related Workmentioning
confidence: 99%
“…The OpenMP threads (workers) require synchronization before requesting and executing chunks, i.e., only the main thread is allowed to call MPI communication functions, such as MPI Send and MPI Receive [14]. Otherwise, a complex implementation is needed to allow individual OpenMP threads to perform MPI calls.…”
Section: Introductionmentioning
confidence: 99%
“…In [17][18][19] new results are presented for loops with dependencies. Recent research results [20,21] have been reported for designing loop self-scheduling methods for grids. In [10,22,23], the heterogeneity of different cluster systems was considered, in order to get better load balancing.…”
Section: Related Workmentioning
confidence: 99%