2021 International Joint Conference on Neural Networks (IJCNN) 2021
DOI: 10.1109/ijcnn52387.2021.9533429
|View full text |Cite
|
Sign up to set email alerts
|

Ripple Walk Training: A Subgraph-based Training Framework for Large and Deep Graph Neural Network

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
38
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
1

Relationship

3
3

Authors

Journals

citations
Cited by 16 publications
(43 citation statements)
references
References 11 publications
0
38
0
Order By: Relevance
“…In such a case, the depth of GNN models can reach more than fifty with a better performance. Till now, the optimized methods based on subgraph learning are limited [14,6,4]. Their measurement of the quality of sampled subgraphs is not comprehensive enough, and the sampling of subgraphs is precisely the key to this type of method.…”
Section: Oversmoothingmentioning
confidence: 99%
See 4 more Smart Citations
“…In such a case, the depth of GNN models can reach more than fifty with a better performance. Till now, the optimized methods based on subgraph learning are limited [14,6,4]. Their measurement of the quality of sampled subgraphs is not comprehensive enough, and the sampling of subgraphs is precisely the key to this type of method.…”
Section: Oversmoothingmentioning
confidence: 99%
“…The learning scheme shown in Equation 1 needs to take the full graph G as input. Training GNN models with the full graph, especially facing large-sized graphs, may easily lead to aforementioned three problems:neighbors explosion, node dependence, and oversmoothing [4]. To solve these problems, subgraph-based training methods [4,14] employ subgraphs of G to construct the mini-batch in each training iteration and update the complete GNN models based on the mini-batch gradient.…”
Section: Subgraph-based Training For Gnn Modelsmentioning
confidence: 99%
See 3 more Smart Citations