Proceedings of the 48th International Conference on Parallel Processing 2019
DOI: 10.1145/3337821.3337862
|View full text |Cite
|
Sign up to set email alerts
|

Distributed Join Algorithms on Multi-CPU Clusters with GPUDirect RDMA

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 37 publications
0
4
0
Order By: Relevance
“…MG-Join: A Scalable Join for Massively Parallel Multi-GPU Architectures We further conduct in-depth experiments to demonstrate the efficiency of MG-Join in joining large data sets. Our experiments find that 1) MG-Join achieves up to 97% utilization of the bisection bandwidth of the GPU interconnect links, and 2) MG-Join achieves up to 2.5x performance improvement over existing multi-GPU hash join implementations[32,212]. Moreover, MG-Join helps improve the overall performance of TPC-H queries by up to 4.5x over multi-GPU version of an open-source commercial GPU database, Omnisci[7].The rest of this chapter is organized as follows.…”
mentioning
confidence: 81%
See 3 more Smart Citations
“…MG-Join: A Scalable Join for Massively Parallel Multi-GPU Architectures We further conduct in-depth experiments to demonstrate the efficiency of MG-Join in joining large data sets. Our experiments find that 1) MG-Join achieves up to 97% utilization of the bisection bandwidth of the GPU interconnect links, and 2) MG-Join achieves up to 2.5x performance improvement over existing multi-GPU hash join implementations[32,212]. Moreover, MG-Join helps improve the overall performance of TPC-H queries by up to 4.5x over multi-GPU version of an open-source commercial GPU database, Omnisci[7].The rest of this chapter is organized as follows.…”
mentioning
confidence: 81%
“…• In Section 5.4.3, we compare the overall performance of MG-Join against UMJ [32] and DPRJ [212]. For communication across GPUs within the same node, DPRJ makes use of CUDA communication APIs that chooses the direct routes between GPUs for data transfer.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations