2017 16th IEEE International Conference on Machine Learning and Applications (ICMLA) 2017
DOI: 10.1109/icmla.2017.00-72
|View full text |Cite
|
Sign up to set email alerts
|

An Investigation of How Neural Networks Learn from the Experiences of Peers Through Periodic Weight Averaging

Abstract: We investigate a method, weighted average model fusion, that enables neural networks to learn from the experiences of other networks, as well as from their own experiences. This method is inspired by the the social natural of humans, which has been shown to be one of the biggest factors in the development of our cognitive abilities. Modern machine learning has focuses predominantly on learning from direct training, and has largely ignored learning through social engagement with peers, neural networks will the … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
5
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(6 citation statements)
references
References 16 publications
1
5
0
Order By: Relevance
“…It was empirically found that vanilla averaging combines the knowledge contained in several neural networks into a single fused model [35,32]. The simple vanilla averaging, however, only works in the case when the weights of individuals networks are relatively close in the weight space.…”
Section: Model Fusionmentioning
confidence: 99%
See 2 more Smart Citations
“…It was empirically found that vanilla averaging combines the knowledge contained in several neural networks into a single fused model [35,32]. The simple vanilla averaging, however, only works in the case when the weights of individuals networks are relatively close in the weight space.…”
Section: Model Fusionmentioning
confidence: 99%
“…The further details of skill transfer are given in Appendix D.1. We change the weight proportion of model B (also known as fusion rate in [32]) to find the weight combination that has the best performance.…”
Section: Skill Transfermentioning
confidence: 99%
See 1 more Smart Citation
“…For similar set of weights of the convolutional part, we cannot use the summation (3) as both terms in this expression would have similar values. Instead, the approach of (Smith & Gashler, 2017;Utans, 1996) would be ideal in this case and we can use weight averaging. However, the top fully connected parts still can be fused by weights summation.…”
Section: Fusion Of Deep Convolutional Neural Networkmentioning
confidence: 99%
“…As discussed in Section 2.4, for WS method we used slightly different procedure as before. As the weights of the convolutional parts of both networks originated from the same set of pretrained weights, during fusion by WS method, those weights were averaged according to approach adopted by Smith & Gashler (2017), Utans (1996). The weights of the classifiers on top of networks were obtained from an independent set of weights and thus were summed up as in all previous examples.…”
Section: Fusion Of Deep Neural Networkmentioning
confidence: 99%