2021
DOI: 10.48550/arxiv.2106.07512
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Last Layer Marginal Likelihood for Invariance Learning

Abstract: Data augmentation is often used to incorporate inductive biases into models. Traditionally, these are hand-crafted and tuned with cross validation. The Bayesian paradigm for model selection provides a path towards end-to-end learning of invariances using only the training data, by optimising the marginal likelihood. We work towards bringing this approach to neural networks by using an architecture with a Gaussian process in the last layer, a model for which the marginal likelihood can be computed. Experimental… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 11 publications
0
2
0
Order By: Relevance
“…In [4] and [8] symmetries are selected using a regularized training loss directly. Many symmetry discovery methods focus on learning invariances [27,4,22,26,10], which are easier to parameterize. This works offers a way to parameterize continuous equivariance constraints, which could allow extensions of symmetry discovery approaches to learnable equivariances.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…In [4] and [8] symmetries are selected using a regularized training loss directly. Many symmetry discovery methods focus on learning invariances [27,4,22,26,10], which are easier to parameterize. This works offers a way to parameterize continuous equivariance constraints, which could allow extensions of symmetry discovery approaches to learnable equivariances.…”
Section: Related Workmentioning
confidence: 99%
“…Secondly, automatically learning symmetry structure from data is an interesting problem. Work in this field which often focuses on invariances [27,4,26,22,10], which are easier to parameterize than equivariances. Parameterizations that allow smooth and adjustable symmetry constraints could extends such methods to layer-by-layer learnable equivariances.…”
Section: Introductionmentioning
confidence: 99%