2021
DOI: 10.48550/arxiv.2103.15419
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Translating Numerical Concepts for PDEs into Neural Architectures

Abstract: We investigate what can be learned from translating numerical algorithms into neural networks. On the numerical side, we consider explicit, accelerated explicit, and implicit schemes for a general higher order nonlinear diffusion equation in 1D, as well as linear multigrid methods. On the neural network side, we identify corresponding concepts in terms of residual networks (ResNets), recurrent networks, and U-nets. These connections guarantee Euclidean stability of specific ResNets with a transposed convolutio… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
10
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(10 citation statements)
references
References 23 publications
0
10
0
Order By: Relevance
“…We note that there is a marginal improvement in the loss at the same time; we show that there is a 3Γ— improvement in training time for a very deep U-Net architecture. This ties into the theme of correlations between U-Net architecture and multigrid methods mentioned in Alt et al [1].…”
Section: Architectural Adaptationmentioning
confidence: 53%
See 3 more Smart Citations
“…We note that there is a marginal improvement in the loss at the same time; we show that there is a 3Γ— improvement in training time for a very deep U-Net architecture. This ties into the theme of correlations between U-Net architecture and multigrid methods mentioned in Alt et al [1].…”
Section: Architectural Adaptationmentioning
confidence: 53%
“…This approximation is plugged into the PDE, after which we invoke Galerkin's method. We multiply the PDE with a test function and reduce the differentiability requirement on 𝑒 β„Ž using integration by parts: 1 ∫…”
Section: Fem Lossmentioning
confidence: 99%
See 2 more Smart Citations
“…We note that there is a marginal improvement in the loss at the same time; we show that there is a 3Γ— improvement in training time for a very deep U-Net architecture. This ties into the theme of correlations between U-Net architecture and multigrid methods mentioned in Alt et al 1 . With both the architectural adaptation and Half-V cycle, in 2D spatial domain with a resolution of 512 Γ— 512, we get a speedup of 3Γ— over the baseline training approach at full resolution.…”
Section: Architectural Adaptationmentioning
confidence: 53%