2008
DOI: 10.1007/978-3-540-68555-5_13
|View full text |Cite
|
Sign up to set email alerts
|

Performance Evaluation of Parallel Sparse Matrix–Vector Products on SGI Altix3700

Abstract: Abstract. The present paper discusses scalable implementations of sparse matrix-vector products, which are crucial for high performance solutions of large-scale linear equations, on a cc-NUMA machine SGI Altix3700. Three storage formats for sparse matrices are evaluated, and scalability is attained by implementations considering the page allocation mechanism of the NUMA machine. Influences of the cache/memory bus architectures on the optimum choice of the storage format are examined, and scalable converters be… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
12
0

Year Published

2008
2008
2020
2020

Publication Types

Select...
3
3
3

Relationship

0
9

Authors

Journals

citations
Cited by 17 publications
(12 citation statements)
references
References 8 publications
0
12
0
Order By: Relevance
“…This operation is considered as a relevant computational kernel in scientific applications, which performs not optimally on modern processors because of the lack of compromise between memory and computing power and irregular memory access patterns ). In general, we find quite a lot of done work in the field of sparse matrix-vector multiplications using parallelization techniques (Kotakemori et al, 2008;Goumas et al, 2009;Williams et al, 2009). These papers study in depth the optimal performance of this operation, but in this paper, we show that even using a simpler parallelization routine, the computation time is noticeably shortened.…”
Section: Related Workmentioning
confidence: 83%
“…This operation is considered as a relevant computational kernel in scientific applications, which performs not optimally on modern processors because of the lack of compromise between memory and computing power and irregular memory access patterns ). In general, we find quite a lot of done work in the field of sparse matrix-vector multiplications using parallelization techniques (Kotakemori et al, 2008;Goumas et al, 2009;Williams et al, 2009). These papers study in depth the optimal performance of this operation, but in this paper, we show that even using a simpler parallelization routine, the computation time is noticeably shortened.…”
Section: Related Workmentioning
confidence: 83%
“…This operation is considered as a relevant computational kernel in scientific applications, which performs not optimally on modern processors because of the lack of compromise between memory and computing power and irregular memory access patterns [5]. In general, we find quite a lot of done work in the field of sparse matrix-vector multiplications using parallelization techniques [6], [7], [8]. These papers study in depth the optimal performance of this operation, but in this paper, we show that even using a simpler parallelization routine, the computation time is noticeably shortened.…”
Section: Related Workmentioning
confidence: 99%
“…In [4], the authors analyze the behavior of the SpMV on a NUMA Altix system, focusing on the effect of using different sparse matrix storage formats. Williams et al [5] propose several optimization techniques for the SpMV which are evaluated on different multicore platforms.…”
Section: Introductionmentioning
confidence: 99%