2024
DOI: 10.1109/tnnls.2022.3223018
|View full text |Cite
|
Sign up to set email alerts
|

On Representation Knowledge Distillation for Graph Neural Networks

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 13 publications
(10 citation statements)
references
References 43 publications
0
10
0
Order By: Relevance
“…We also demonstrate that, in general, the use of a projector will scales much more favourably for larger batch sizes and feature dimensions. We also note that the handcrafted design of kernel functions [28,21,38] may not generalise to large scale or complex real-world datasets. From the results in table 2, we observe that when fixing all other settings, the choice of normalisation can significantly affect the student's performance.…”
Section: Revisiting Knowledge Distillationmentioning
confidence: 99%
“…We also demonstrate that, in general, the use of a projector will scales much more favourably for larger batch sizes and feature dimensions. We also note that the handcrafted design of kernel functions [28,21,38] may not generalise to large scale or complex real-world datasets. From the results in table 2, we observe that when fixing all other settings, the choice of normalisation can significantly affect the student's performance.…”
Section: Revisiting Knowledge Distillationmentioning
confidence: 99%
“…GFKD [46] RDD [47] GKD [48] GLNN [49] Distill2Vec [50] MT-GCN [51] TinyGNN [52] GLocalKD [53] SCR [54] ROD [55] EGNN [56] Middle layer LWC-KD [57] MustaD [58] EGAD [59] AGNN [60] Cold Brew [61] PGD [62] OAD [63] CKD [64] BGNN [65] EGSC [66] HSKDM [67] Constructed graph GRL [68] GFL [69] HGKT [70] CPF [71] LSP [16] scGCN [72] MetaHG [73] G-CRD [74] HIRE [75] SKD methods…”
Section: Output Layermentioning
confidence: 99%
“…To further enrich and provide more general knowledge, GNNs will further learn the topological structure and node relationship information of the teacher model with the help of constructed graphs [16,[68][69][70][71][72][73][74][75], so as to deeply explore the knowledge contained in the teacher model.…”
Section: Constructed Graph Knowledgementioning
confidence: 99%
See 1 more Smart Citation
“…We also investigated a number of different transformation functions and the effect they have on performance. The transformations we utilized include: the identity transformation (when appropriate); learned linear transformations and MLP projection heads; as well as variants of the structure-preserving transformations described in [36]. Our results (see Appendix E.1) showed that the examined transformations either hurt performance or only provided a marginal improvement in performance compared to our default linear mapping.…”
Section: Effect Of Transformation Functionmentioning
confidence: 99%