2008
DOI: 10.1007/s11227-008-0251-8
|View full text |Cite
|
Sign up to set email alerts
|

Performance evaluation of the sparse matrix-vector multiplication on modern architectures

Abstract: In this paper, we revisit the performance issues of the widely used sparse matrix-vector multiplication (SpMxV) kernel on modern microarchitectures. Previous scientific work reports a number of different factors that may significantly reduce performance. However, the interaction of these factors with the underlying architectural characteristics is not clearly understood, a fact that may lead to misguided, and thus unsuccessful attempts for optimization. In order to gain an insight into the details of SpMxV per… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
52
0

Year Published

2010
2010
2022
2022

Publication Types

Select...
4
4
2

Relationship

2
8

Authors

Journals

citations
Cited by 86 publications
(54 citation statements)
references
References 23 publications
0
52
0
Order By: Relevance
“…This operation is considered as a relevant computational kernel in scientific applications, which performs not optimally on modern processors because of the lack of compromise between memory and computing power and irregular memory access patterns ). In general, we find quite a lot of done work in the field of sparse matrix-vector multiplications using parallelization techniques (Kotakemori et al, 2008;Goumas et al, 2009;Williams et al, 2009). These papers study in depth the optimal performance of this operation, but in this paper, we show that even using a simpler parallelization routine, the computation time is noticeably shortened.…”
Section: Related Workmentioning
confidence: 83%
“…This operation is considered as a relevant computational kernel in scientific applications, which performs not optimally on modern processors because of the lack of compromise between memory and computing power and irregular memory access patterns ). In general, we find quite a lot of done work in the field of sparse matrix-vector multiplications using parallelization techniques (Kotakemori et al, 2008;Goumas et al, 2009;Williams et al, 2009). These papers study in depth the optimal performance of this operation, but in this paper, we show that even using a simpler parallelization routine, the computation time is noticeably shortened.…”
Section: Related Workmentioning
confidence: 83%
“…This operation is considered as a relevant computational kernel in scientific applications, which performs not optimally on modern processors because of the lack of compromise between memory and computing power and irregular memory access patterns [5]. In general, we find quite a lot of done work in the field of sparse matrix-vector multiplications using parallelization techniques [6], [7], [8]. These papers study in depth the optimal performance of this operation, but in this paper, we show that even using a simpler parallelization routine, the computation time is noticeably shortened.…”
Section: Related Workmentioning
confidence: 99%
“…Past work [12,2] has identified the memory subsystem as the main performance bottleneck of the SpMV kernel. Obviously, this problem becomes more severe in a multithreaded environment, where multiple cores access the main memory.…”
Section: Specialization On User Inputmentioning
confidence: 99%