2019
DOI: 10.48550/arxiv.1909.03211
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Measuring and Relieving the Over-smoothing Problem for Graph Neural Networks from the Topological View

Abstract: Graph Neural Networks (GNNs) have achieved promising performance on a wide range of graph-based tasks. Despite their success, one severe limitation of GNNs is the over-smoothing issue (indistinguishable representations of nodes in different classes). In this work, we present a systematic and quantitative study on the over-smoothing issue of GNNs. First, we introduce two quantitative metrics, MAD and MADGap, to measure the smoothness and oversmoothness of the graph nodes representations, respectively. Then, we … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
28
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 17 publications
(30 citation statements)
references
References 22 publications
2
28
0
Order By: Relevance
“…(2) The performance improvements of AdaGNN-R and AdaGNN-S are more obvious on BlogCatalog and Flickr compared with other datasets. The reason could be attributed to their high average node degree of social network (as indicated in Table 1) -nodes are influenced by more neighbors during neighborhood aggregation, which is consistent with the observations in previous literature [6]. For such dataset with high average node degree, the adaptive frequency response provided by AdaGNN can help achieve more appropriate feature smoothness.…”
Section: Resultssupporting
confidence: 82%
See 1 more Smart Citation
“…(2) The performance improvements of AdaGNN-R and AdaGNN-S are more obvious on BlogCatalog and Flickr compared with other datasets. The reason could be attributed to their high average node degree of social network (as indicated in Table 1) -nodes are influenced by more neighbors during neighborhood aggregation, which is consistent with the observations in previous literature [6]. For such dataset with high average node degree, the adaptive frequency response provided by AdaGNN can help achieve more appropriate feature smoothness.…”
Section: Resultssupporting
confidence: 82%
“…Li et al [29] proved that GCN is actually a kind of Laplacian smoothing process, and proposed the challenge of oversmoothing for the first time. After that, some studies demonstrate that certain level of smoothness benefits node representation learning while over-smoothing broadly exists in deeper GNNs [6,12]. More recently, some researches attempt to relieve this problem via residual-like connections [8,31,32].…”
Section: Resultsmentioning
confidence: 99%
“…Node pair distance has been widely adopted to quantify the over-smoothing based on embedding similarities [18,22]. Among the series of distance metrics, Dirichlet energy is simple and expressive for the over-smoothing analysis [32].…”
Section: Dirichlet Energy Constrained Learningmentioning
confidence: 99%
“…As the layer number increases, the node representations will converge to indistinguishable vectors due to the recursive neighborhood aggregation and non-linear activation [15,16]. Such phenomenon is recognized as over-smoothing issue [17,18,19,20,21]. It prevents the stacking of many layers and modeling the dependencies to high-order neighbors.…”
Section: Introductionmentioning
confidence: 99%
“…This, however, is not always successful in capturing information from far-away nodes as information can be aggregated from too many nodes, drowning out relevant contributions. This is the over-smoothing phenomenon [18,19] which, for excessively large aggregation ranges, can produce output features that are very similar across the different nodes since the aggregation ranges of the different nodes overlap strongly.…”
Section: Related Workmentioning
confidence: 99%