2022
DOI: 10.1109/tcad.2021.3079142
|View full text |Cite
|
Sign up to set email alerts
|

Rubik: A Hierarchical Architecture for Efficient Graph Neural Network Training

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
6
3

Relationship

1
8

Authors

Journals

citations
Cited by 30 publications
(7 citation statements)
references
References 34 publications
0
7
0
Order By: Relevance
“…The current taxonomy captures the intra-phase dataflows and the inter-phase dataflows. However, our taxonomy does not capture the order of nodes, graph partitioning and optimizations such as load balancing [20], computation elimination via memoizing [23], [36] and requires an extension to capture these.…”
Section: Discussion and Future Workmentioning
confidence: 99%
“…The current taxonomy captures the intra-phase dataflows and the inter-phase dataflows. However, our taxonomy does not capture the order of nodes, graph partitioning and optimizations such as load balancing [20], computation elimination via memoizing [23], [36] and requires an extension to capture these.…”
Section: Discussion and Future Workmentioning
confidence: 99%
“…An increasing amount of research dealing with the co-design of software and hardware to accelerate GNN training [23,95,96,203]. Here, not only software and algorithms are optimized, but also hardware modules are developed to better address the characteristics of GNNs.…”
Section: Current Research Trendsmentioning
confidence: 99%
“…Some of the works like HyGCN [69], AWB [25], and VersaGNN [60] making effort on sparse matrix multiplication, but it is not the workflow for LSD-GNN. Others such as GCNAX [44], BoostGCN [77], Rubik [14], GraphACT [76], GNNSampler [47], and Grip [39,40] optimizing data reuse are not applicable to LSD-GNN either, since the chance to find reuse within 512-node mini-batch compared with 10+ billion total nodes is extremely low. Huang et al [34] works on a similar problem like this paper, but under a different disaggregated memory pool context.…”
Section: Related Workmentioning
confidence: 99%