2021
DOI: 10.48550/arxiv.2103.14686
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Generalization capabilities of translationally equivariant neural networks

Srinath Bulusu,
Matteo Favoni,
Andreas Ipp
et al.
Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 39 publications
0
3
0
Order By: Relevance
“…The target densities defined in Table I are all invariant under translations with appropriate boundary conditions, as discussed in Section II C. Previous works have shown that exactly incorporating known symmetries into machine learning models can accelerate their training and improve their final quality [86][87][88][89][90][91]. In the context of normalizing flows, ensuring that the model density is invariant under a symmetry group is achieved by choosing an invariant prior distribution and building transformation layers that are equivariant under the symmetry.…”
Section: B Building Blocksmentioning
confidence: 99%
“…The target densities defined in Table I are all invariant under translations with appropriate boundary conditions, as discussed in Section II C. Previous works have shown that exactly incorporating known symmetries into machine learning models can accelerate their training and improve their final quality [86][87][88][89][90][91]. In the context of normalizing flows, ensuring that the model density is invariant under a symmetry group is achieved by choosing an invariant prior distribution and building transformation layers that are equivariant under the symmetry.…”
Section: B Building Blocksmentioning
confidence: 99%
“…Our results prove that linear G-CNNs are biased towards low rank solutions in the Fourier regime, via regularization of the 2/L-Schatten norms over Fourier matrix coefficients (also a quasi-norm). Lastly, there is a line of work focusing on understanding the expressivity Cohen et al, 2019;Yarotsky, 2021) and generalization (Sannai & Imaizumi, 2019;Lyle et al, 2020;Bulusu et al, 2021;Elesedy & Zaidi, 2021) of equivariant networks but not specifically the effects of implicit regularization.…”
Section: Related Workmentioning
confidence: 99%
“…Machine learning (ML) techniques provide powerful algorithms building models based on the correlation pattern of data. Various approaches enhancing the lattice QCD calculations using ML have been explored in recent studies [1][2][3][4]. However, the ML training, which is the procedure of finding optimal model parameters describing the data, of most of the ML algorithms involve nontrivial optimization problems.…”
Section: Introductionmentioning
confidence: 99%