2008
DOI: 10.1145/1377603.1377606
|View full text |Cite
|
Sign up to set email alerts
|

Families of algorithms related to the inversion of a Symmetric Positive Definite matrix

Abstract: We study the high-performance implementation of the inversion of a Symmetric Positive Definite (SPD) matrix on architectures ranging from sequential processors to Symmetric MultiProcessors to distributed memory parallel computers. This inversion is traditionally accomplished in three "sweeps": a Cholesky factorization of the SPD matrix, the inversion of the resulting triangular matrix, and finally the multiplication of the inverted triangular matrix by its own transpose. We state different algorithms for each … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
42
0

Year Published

2011
2011
2021
2021

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 54 publications
(42 citation statements)
references
References 27 publications
0
42
0
Order By: Relevance
“…However, libflame differs from LAPACK in two impor-tant ways. First, as mentioned, it provides families of algorithms for each operation so that the best can be chosen for a given circumstance [3]. Second, it provides a framework for building complete custom linear algebra codes.…”
Section: What Is Differentmentioning
confidence: 99%
See 1 more Smart Citation
“…However, libflame differs from LAPACK in two impor-tant ways. First, as mentioned, it provides families of algorithms for each operation so that the best can be chosen for a given circumstance [3]. Second, it provides a framework for building complete custom linear algebra codes.…”
Section: What Is Differentmentioning
confidence: 99%
“…In our publications and performance graphs, we do our best to dispel the myth that user-and programmer-friendly linear algebra codes cannot yield high performance. Our FLAME implementations of operations such as Cholesky factorization and triangular matrix inversion often outperform the corresponding implementations currently available in LAPACK [3]. In Section 3 and numerous papers, we give numerous examples of how performance is facilitated by the FLAME methodology.…”
Section: What Is Differentmentioning
confidence: 99%
“…For other operations, targeting sequential, multithreaded, and distributed memory architectures, we have consistently shown the benefit of having a choice of algorithms. Several illuminating examples, including derivations and performance comparisons, can be found in [GGHvdG01, BGM + 05, QOvdG03,vdGQO08,BGdG]. …”
Section: Performancementioning
confidence: 99%
“…Moreover, as part of the FLAME project we have long advocated that it is important to have multiple algorithmic variants at our disposal so that the best algorithm can be chosen for each situation [4]. The FLAME methodology advocates systematic derivation of these variants [3,11].…”
Section: Accelerating the Cublasmentioning
confidence: 99%