2016 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW) 2016
DOI: 10.1109/ipdpsw.2016.41
|View full text |Cite
|
Sign up to set email alerts
|

A Comparison of High-Level Programming Choices for Incomplete Sparse Factorization Across Different Architectures

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
9
0

Year Published

2019
2019
2019
2019

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(9 citation statements)
references
References 29 publications
0
9
0
Order By: Relevance
“…Incomplete levels. We will only consider ILU(k) with k = 0, though Javelin supports other levels as implemented by other work [5], [6], [15] and commonly used in iterative solvers. As k increases, additional fill-in is allowed into the sparsity pattern.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Incomplete levels. We will only consider ILU(k) with k = 0, though Javelin supports other levels as implemented by other work [5], [6], [15] and commonly used in iterative solvers. As k increases, additional fill-in is allowed into the sparsity pattern.…”
Section: Methodsmentioning
confidence: 99%
“…Therefore, comparing performance as k varies would not provide a deep understanding of the scalability of Javelin without knowing where and how much fill-in was produced. It is therefore common to compare scalability primarily with ILU(0) [3], [5] over a large test suite of matrices with different sparsity pattern and row density in order to estimate how well the implementation scales with fill-in as we do in this paper.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Parallel programs should be optimised to extract maximum performance from hardware on architecture case by case [38], which is far from trivial according to Booth et al [39]. There exist different and combined manners to explore parallelism such as Data Parallelism and Task Parallelism [40].…”
Section: Introductionmentioning
confidence: 99%
“…Parallel programs should be optimized to extract maximum performance from hardware on architecture case by case, which is far from trivial according to (Booth et al, 2016). There exist different and combined manners to explore parallelism such as data parallelism and task parallelism (Gordon et al, 2006).…”
Section: Introductionmentioning
confidence: 99%