2021
DOI: 10.48550/arxiv.2112.15345
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Distributed Hybrid CPU and GPU training for Graph Neural Networks on Billion-Scale Graphs

Abstract: Graph neural networks (GNN) have shown great success in learning from graph-structured data. They are widely used in various applications, such as recommendation, fraud detection, and search. In these domains, the graphs are typically large, containing hundreds of millions or billions of nodes. To tackle this challenge, we develop DistDGLv2, a system that extends DistDGL for training GNNs in a mini-batch fashion, using distributed hybrid CPU/GPU training to scale to large graphs. DistDGLv2 places graph data in… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
9
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(10 citation statements)
references
References 15 publications
(23 reference statements)
0
9
0
Order By: Relevance
“…Representations of large graphs can exceed CPU memory and either require distribution to multiple machines [12,19,37,40,49,50] or storing graph data on disk [22,29]. In disk-based training the node representations are divided into partitions that are swapped in-and-out of CPU memory.…”
Section: Scaling Gnn Training To Compute ℎ (𝑘)mentioning
confidence: 99%
See 4 more Smart Citations
“…Representations of large graphs can exceed CPU memory and either require distribution to multiple machines [12,19,37,40,49,50] or storing graph data on disk [22,29]. In disk-based training the node representations are divided into partitions that are swapped in-and-out of CPU memory.…”
Section: Scaling Gnn Training To Compute ℎ (𝑘)mentioning
confidence: 99%
“…Large-Scale Training To scale GNN training to graphs which exceed the CPU memory capacity of a single box, many works opt for a distributed multi-machine approach [12,19,49]. In particular, recent work introduces DistDGLv2 as a distributed version of DGL [49].…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations