2015
DOI: 10.1016/j.parco.2015.06.004
|View full text |Cite
|
Sign up to set email alerts
|

Blocking and parallelization of the Hari–Zimmermann variant of the Falk–Langemeyer algorithm for the generalized SVD

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
22
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
7

Relationship

1
6

Authors

Journals

citations
Cited by 19 publications
(22 citation statements)
references
References 27 publications
0
22
0
Order By: Relevance
“…If A and B are instead given implicitly by their factors F and G (not necessarily square nor with the same number of rows), respectively, such that ( A , B ) = ( F F , G G ) , then the GEVD of ( A , B ) can be computed implicitly, i.e. without assembling A and B in entirety from the factors, by a modification of the Hari–Zimmermann algorithm (Novaković et al, 2015). However, pivot submatrices of A and B of a certain, usually small order are formed explicitly throughout the computation.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…If A and B are instead given implicitly by their factors F and G (not necessarily square nor with the same number of rows), respectively, such that ( A , B ) = ( F F , G G ) , then the GEVD of ( A , B ) can be computed implicitly, i.e. without assembling A and B in entirety from the factors, by a modification of the Hari–Zimmermann algorithm (Novaković et al, 2015). However, pivot submatrices of A and B of a certain, usually small order are formed explicitly throughout the computation.…”
Section: Introductionmentioning
confidence: 99%
“…The recent work (Novaković et al, 2015) has shown that such method can be blocked and parallelized for the shared memory nodes and for the clusters of those, albeit only the real matrix pairs were considered therein. Even the sequential but blocked version outperformed the GSVD algorithm in LAPACK (Anderson et al, 1999), and the parallel ones exhibited a decent scalability.…”
Section: Introductionmentioning
confidence: 99%
“…Nonetheless, the iteration matrices occasionally need normalization during the process. When the FL method is used as a kernel algorithm for the block Jacobi method, this can be a demanding task on contemporary CPU and GPU parallel computing machines [6]. Fortunately, one can use other Jacobi methods as kernel algorithms, in particular the Hari-Zimmermann (HZ) and the Cholesky-Jacobi (CJ) method (see [1,8]) which can be seen as normalized versions of the FL method.…”
Section: Introductionmentioning
confidence: 99%
“…This typically happens in the course of modelling the parameters of a system. The described properties make the method an excellent choice for the kernel algorithm of the block-Jacobi methods which are nowadays the prime choice of the methods for solving the definite GEP on contemporary parallel CPU and GPU computing machines [7].In this short communication, we present the main formulas of the complex Falk-Langemeyer (CFL) algorithm that are derived in [3]. Although they are the proper generalization of those in the real method, their derivation is not trivial.…”
mentioning
confidence: 99%
“…This typically happens in the course of modelling the parameters of a system. The described properties make the method an excellent choice for the kernel algorithm of the block-Jacobi methods which are nowadays the prime choice of the methods for solving the definite GEP on contemporary parallel CPU and GPU computing machines [7].…”
mentioning
confidence: 99%