2018
DOI: 10.1016/j.parco.2017.12.004
|View full text |Cite
|
Sign up to set email alerts
|

A distributed-memory hierarchical solver for general sparse linear systems

Abstract: We present a parallel hierarchical solver for general sparse linear systems on distributed-memory machines. For large-scale problems, this fully algebraic algorithm is faster and more memory-efficient than sparse direct solvers because it exploits the low-rank structure of fill-in blocks. Depending on the accuracy of low-rank approximations, the hierarchical solver can be used either as a direct solver or as a preconditioner. The parallel algorithm is based on data decomposition and requires only local communi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
27
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
6
2

Relationship

4
4

Authors

Journals

citations
Cited by 23 publications
(27 citation statements)
references
References 41 publications
0
27
0
Order By: Relevance
“…This parallelization could resolve the issue of huge memory consumption, which we experienced on shared-memory systems. For this regard, the parallelization proposed for a sparse direct solver (LoRaSp [23]) by Chen et al [22] would be helpful. However, the modification to dense systems is not trivial because the algorithm of IFMM is significantly different from that of LoRaSp.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…This parallelization could resolve the issue of huge memory consumption, which we experienced on shared-memory systems. For this regard, the parallelization proposed for a sparse direct solver (LoRaSp [23]) by Chen et al [22] would be helpful. However, the modification to dense systems is not trivial because the algorithm of IFMM is significantly different from that of LoRaSp.…”
Section: Resultsmentioning
confidence: 99%
“…However, this overfine-grained parallelization would inevitably introduce overwhelming overhead when the number of threads is large. Another closely related approach is the distributed-memory parallel solver introduced by Chen et al [22], which parallelizes the LoRaSp algorithm [23], an analog of the IFMM for solving 1 when A is sparse.…”
Section: Introductionmentioning
confidence: 99%
“…Sparsification would then have to be ordered by color, i.e., s i is sparsified before s j if and only if c i < c j . This is similar to what is done in the parallel version of LoRaSp, see [11]. As a result, we used Algorithm 4 to maximize concurrency.…”
Section: Simultaneous Sparsificationmentioning
confidence: 96%
“…Different works propose efficient parallel solvers of sparse linear systems: novel strategies and related performance figures are compared with wellknown software packages, together with the discussion of theoretical complexity and accuracy, and/or the use of specific parallel paradigms, run-time systems and accelerators. In [13], a parallel hierarchical solver is proposed and compared with the SuperLU performances. In [39], the performances of a novel multi-frontal solver are discussed and compared with MUMPS [5] and SuperLU.…”
Section: Related Work Novelty and Main Contributionsmentioning
confidence: 99%