2020
DOI: 10.1016/j.neunet.2020.03.016
|View full text |Cite
|
Sign up to set email alerts
|

Learning in the machine: To share or not to share?

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
2

Relationship

2
7

Authors

Journals

citations
Cited by 14 publications
(9 citation statements)
references
References 29 publications
0
8
0
Order By: Relevance
“…(2020) , which may be applied in future studies, and generally makes it more comparable to that work. Second, because of the weight-sharing effects of convolutions, which reduce the total number of trainable parameters, it acts as a regularization method that builds on the dropout and batch normalization layers, which is appropriate for small training samples that may be prone to overfitting ( Kukačka, Golkov, Cremers, 2017 , Ott, Linstead, LaHaye, Baldi, 2020 ). To further validate the model, we additionally compare it to the performance of an ensemble of fully-connected neural networks lacking the convolutional layers.…”
Section: Methodsmentioning
confidence: 99%
“…(2020) , which may be applied in future studies, and generally makes it more comparable to that work. Second, because of the weight-sharing effects of convolutions, which reduce the total number of trainable parameters, it acts as a regularization method that builds on the dropout and batch normalization layers, which is appropriate for small training samples that may be prone to overfitting ( Kukačka, Golkov, Cremers, 2017 , Ott, Linstead, LaHaye, Baldi, 2020 ). To further validate the model, we additionally compare it to the performance of an ensemble of fully-connected neural networks lacking the convolutional layers.…”
Section: Methodsmentioning
confidence: 99%
“…This is easily achieved through data augmentation by translating each training image in all possible directions, something that may happen automatically in the real world due to moving objects, or head/eye motions. With this data augmentation, the weights of the convolution neurons remain similar throughout training, since they are trained on the same data, without any exact weight sharing [Ott et al, 2020]. This approach ought to be tried in Tourbillon.…”
Section: Discussionmentioning
confidence: 99%
“…We can think of receptive field based neurons organized in a hierarchical architecture that carry out translation equivariance without sharing their weights. This is strongly motivated also by the arguable biological plausibility of the mechanism of weight sharing [76]. Such a lack of plausibility is more serious than the supposed lack of a truly local computational scheme in Backpropagation, which mostly comes from the lack of delay in the forward model of the neurons [14].…”
Section: Why Receptive Fields and Hierarchical Architectures?mentioning
confidence: 99%