2022
DOI: 10.1007/s11227-022-04932-3
|View full text |Cite
|
Sign up to set email alerts
|

NAS Parallel Benchmarks with Python: a performance and programming effort analysis focusing on GPUs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
3
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(4 citation statements)
references
References 22 publications
1
3
0
Order By: Relevance
“…The performance gap widened with increasing N and reached about 22x and 7.8x compared to Numba and CuPy for N = 2 × 10 9 . These findings are in line with those of previous studies (e.g., [39,41]).…”
Section: Resultssupporting
confidence: 94%
See 3 more Smart Citations
“…The performance gap widened with increasing N and reached about 22x and 7.8x compared to Numba and CuPy for N = 2 × 10 9 . These findings are in line with those of previous studies (e.g., [39,41]).…”
Section: Resultssupporting
confidence: 94%
“…The generated PRNs are immediately consumed within the kernel without storing them in the global memory. We compared our results to our implementation in CUDA C. Note that the 1D MCRT test problem was implemented with 15, 26, and 37 lines of code in CuPy, Numba, and CUDA C, respectively, reflecting the ease of implementation in CuPy and Numba relative to CUDA C [41].…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations