2013
DOI: 10.1137/110848803
|View full text |Cite
|
Sign up to set email alerts
|

High-Performance Solvers for Dense Hermitian Eigenproblems

Abstract: We introduce a new collection of solvers-subsequently called EleMRRR-for largescale dense Hermitian eigenproblems. EleMRRR solves various types of problems: generalized, standard, and tridiagonal eigenproblems. Among these, the last is of particular importance as it is a solver in its own right, as well as the computational kernel for the first two; we present a fast and scalable tridiagonal solver based on the algorithm of multiple relatively robust representations-referred to as PMRRR. Like the other EleMRRR… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

2
35
0

Year Published

2013
2013
2018
2018

Publication Types

Select...
6
1
1

Relationship

1
7

Authors

Journals

citations
Cited by 24 publications
(37 citation statements)
references
References 47 publications
2
35
0
Order By: Relevance
“…On the other hand, the algorithm employed for the solution of the tridiagonal problem, (IV), is a rather new parallel implementation of the MRRR algorithm by Petschow and coworkers. [24] A recent detailed benchmark compared components from ScaLAPACK, Elemental, and the ELPA library on the BlueGene/Q architecture [73], including the important aspect of hyperthreading (the possible use of more than one MPI task per CPU core) on that machine. In short, ELPA and Elemental showed the best performance on this architecture, with a slight edge for the two-step reduction approach (IIIa) and (IIIb) as implemented in ELPA.…”
Section: Recent Developments In Parallel Eigenvalue Solversmentioning
confidence: 99%
“…On the other hand, the algorithm employed for the solution of the tridiagonal problem, (IV), is a rather new parallel implementation of the MRRR algorithm by Petschow and coworkers. [24] A recent detailed benchmark compared components from ScaLAPACK, Elemental, and the ELPA library on the BlueGene/Q architecture [73], including the important aspect of hyperthreading (the possible use of more than one MPI task per CPU core) on that machine. In short, ELPA and Elemental showed the best performance on this architecture, with a slight edge for the two-step reduction approach (IIIa) and (IIIb) as implemented in ELPA.…”
Section: Recent Developments In Parallel Eigenvalue Solversmentioning
confidence: 99%
“…The study is limited to sequential executions and does not take into account the degree of parallelism the algorithms provide. However, various studies [4,38,33,35,32] of the performance and accuracy of parallel implementations come to similar conclusions.…”
Section: Introductionmentioning
confidence: 80%
“…12 Criterion I is used in LAPACK [12] and in results of mr3smp in [32], which usually uses II. Criterion II is used in ScaLAPACK [38] and Elemental [33]. In massively parallel computing environments, criteria III and IV can (and should) be additionally complemented with the splitting based on absolute gaps; see also [34].…”
Section: Adjusting the Algorithmmentioning
confidence: 99%
“…In addition, the highly scalable parallel implementation of the MRRR algorithm is also proposed [15]. However, for the matrices that have extremely clustered eigenvalues such as the Glued-Wilkinson matrix [16], [17], the execution time for the MRRR algorithm is reported to be longer than the time for the bisection algorithm and the inverse iteration algorithm, even though the MRRR is superior with respect to the computational cost [16].…”
Section: Introductionmentioning
confidence: 99%