2022
DOI: 10.1016/j.patter.2022.100491
|View full text |Cite
|
Sign up to set email alerts
|

Scalable deeper graph neural networks for high-performance materials property prediction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
42
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 46 publications
(51 citation statements)
references
References 48 publications
0
42
0
Order By: Relevance
“…This architecture was then used for all models. Interestingly while past works 40 have found the CGCNN can scale up to 25 graph convolutional layers we found that models with more than eight hidden layers suffered from the vanishing gradient problem 41 when training on the relaxed data while training on the augmented data allowed for deeper models. we speculate the models' ability to scale with the number of graph convolutional layers and not the number of hidden layers is a product of the graph convolutional layers containing batch normalization.…”
Section: Trainingmentioning
confidence: 71%
“…This architecture was then used for all models. Interestingly while past works 40 have found the CGCNN can scale up to 25 graph convolutional layers we found that models with more than eight hidden layers suffered from the vanishing gradient problem 41 when training on the relaxed data while training on the augmented data allowed for deeper models. we speculate the models' ability to scale with the number of graph convolutional layers and not the number of hidden layers is a product of the graph convolutional layers containing batch normalization.…”
Section: Trainingmentioning
confidence: 71%
“…To learn the sophisticated structure to property relationship between the crystals and their vibrational frequency, we use our recently developed scalable deeper graph neural networks with a global attention mechanism. 32 Our deeperGATGNN model ( Figure 2 ) is composed of a set of augmented graph attention layers with ResNet style skip connections and differentiable group normalization to achieve complex deep feature extractions. After several such feature transformation steps, a global attention layer is used to aggregate the features at all nodes and a global pooling operator is further used to process the information to generate a latent feature representation for the crystal.…”
Section: Methodsmentioning
confidence: 99%
“…It is composed of several graph convolution layers with differentiable normalization and skip connections plus a global attention layer and final fully connected layers. Reproduced with permission from ref ( 32 ). Copyright 2022 Elsevier (in Patterns ).…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations