2021
DOI: 10.48550/arxiv.2102.10333
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Provably Strict Generalisation Benefit for Equivariant Models

Abstract: It is widely believed that engineering a model to be invariant/equivariant improves generalisation.Despite the growing popularity of this approach, a precise characterisation of the generalisation benefit is lacking. By considering the simplest case of linear models, this paper provides the first provably non-zero improvement in generalisation for invariant/equivariant models when the target distribution is invariant/equivariant with respect to a compact group. Moreover, our work reveals an interesting relatio… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
40
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 9 publications
(40 citation statements)
references
References 17 publications
0
40
0
Order By: Relevance
“…While their framework is potentially applicable to general groups, their results only cover the case of translation groups, leading to improvements in sample complexity of at most a factor equal to the dimension. [15] also sudies benefits of group invariance, but focuses on linear models, and only considers interpolating estimators. [10] study benefits of equivariant kernels in structured prediction problems.…”
Section: Related Workmentioning
confidence: 99%
“…While their framework is potentially applicable to general groups, their results only cover the case of translation groups, leading to improvements in sample complexity of at most a factor equal to the dimension. [15] also sudies benefits of group invariance, but focuses on linear models, and only considers interpolating estimators. [10] study benefits of equivariant kernels in structured prediction problems.…”
Section: Related Workmentioning
confidence: 99%
“…Assumptions and technical conditions are given in Section 2 along with an outline of the ideas of Elesedy and Zaidi [8] on which we build. Related works are discussed in Section 5.…”
Section: More Specifically Letmentioning
confidence: 99%
“…However, while implementations and practical applications abound, until very recently a rigorous theoretical justification for invariance was missing. As pointed out in [8], many prior works such as [29,25] provide only worst-case guarantees on the performance of invariant algorithms. It follows that these results do not rule out the possibility of modern training algorithms automatically favouring invariant models, irrespective of the choice of architecture.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…In other words, it is not the symmetry of X or Y that is relevant, but the symmetry of the function f mapping from X to Y . If the true relationship in the data has a symmetry, then constraining the hypothesis space to functions f that also have the symmetry makes learning easier and improves generalization [15]. Equivariant models have been developed for a wide variety of symmetries and data types like images [10,58,61,56], sets [60,39], graphs [38], point clouds [3,18,47], dynamical systems [16], jets [6], and other objects [54,17].…”
Section: Representationsmentioning
confidence: 99%