2017 IEEE International Conference on Microwaves, Antennas, Communications and Electronic Systems (COMCAS) 2017
DOI: 10.1109/comcas.2017.8244769
|View full text |Cite
|
Sign up to set email alerts
|

Acceleration of in-core LU-decomposition of dense MoM matrix by parallel usage of multiple GPUs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(1 citation statement)
references
References 6 publications
0
1
0
Order By: Relevance
“…Zhou (2016) [5] performed LU decomposition on huge and complex matrices based on the idea of parallel computing and transformed three optimal parallel computing processing methods using three parallel modes (OpenMP, MPI, PPL). Branko et al (2017) [6] proposed using multiple GPUs in parallel to accelerate in-kernel LU decomposition and using blocks to overcome the memory limitations of GPUs. Volkov et al (2009) [7] implemented right-looking algorithms for LU, Cholesky, and QR for GPUs, which are similar to those in the GPU-based advanced linear algebra library MAGMA (2014) [8], changing the data layout by transposing matrices.…”
Section: Introductionmentioning
confidence: 99%
“…Zhou (2016) [5] performed LU decomposition on huge and complex matrices based on the idea of parallel computing and transformed three optimal parallel computing processing methods using three parallel modes (OpenMP, MPI, PPL). Branko et al (2017) [6] proposed using multiple GPUs in parallel to accelerate in-kernel LU decomposition and using blocks to overcome the memory limitations of GPUs. Volkov et al (2009) [7] implemented right-looking algorithms for LU, Cholesky, and QR for GPUs, which are similar to those in the GPU-based advanced linear algebra library MAGMA (2014) [8], changing the data layout by transposing matrices.…”
Section: Introductionmentioning
confidence: 99%