2019
DOI: 10.48550/arxiv.1911.13299
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

What's Hidden in a Randomly Weighted Neural Network?

Abstract: Training a neural network is synonymous with learning the values of the weights. In contrast, we demonstrate that randomly weighted neural networks contain subnetworks which achieve impressive performance without ever training the weight values. Hidden in a randomly weighted Wide ResNet-50 [28] we show that there is a subnetwork (with random weights) that is smaller than, but matches the performance of a ResNet-34 [8] trained on ImageNet [3]. Not only do these "untrained subnetworks" exist, but we provide an a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

2
31
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
7

Relationship

1
6

Authors

Journals

citations
Cited by 10 publications
(33 citation statements)
references
References 11 publications
2
31
0
Order By: Relevance
“…The hypothesis space is composed of all candidate networks with different assignments of weight values that accomplish the computation task. Consistent with previous studies [7][8][9], the optimal random network ensemble includes sub-networks of the original full network, which further allows for capturing uncertainty in the hypothesis space. The model can be solved by mean-field methods, thereby providing a physics interpretation of how credit assignment occurs in a hierarchical deep neural system.…”
supporting
confidence: 55%
See 2 more Smart Citations
“…The hypothesis space is composed of all candidate networks with different assignments of weight values that accomplish the computation task. Consistent with previous studies [7][8][9], the optimal random network ensemble includes sub-networks of the original full network, which further allows for capturing uncertainty in the hypothesis space. The model can be solved by mean-field methods, thereby providing a physics interpretation of how credit assignment occurs in a hierarchical deep neural system.…”
supporting
confidence: 55%
“…Excitingly, recent works showed that there exist subnetworks of random weights that are able to produce better-than-chance accuracies [7][8][9]. This property seems to be universal across different architectures, datasets and computational tasks [10].…”
mentioning
confidence: 95%
See 1 more Smart Citation
“…We leverage this to develop a flexible model capable of learning thousands of tasks: Supermasks in Superposition (SupSup). SupSup, diagrammed in Figure 1, is driven by two core ideas: a) the expressive power of untrained, randomly weighted subnetworks [55,38], and b) inference of task-identity as a gradient-based optimization problem.…”
Section: Introductionmentioning
confidence: 99%
“…• Fact 6: Terms in (13), wherein, the base case is generated as • Fact 8: For k crossings of the base paths there are 4 k+1 splicings possible, and those many terms are extra in the E K 0 (s, s ) 2 expression in (13), when compared to the E [K 0 (s, s )] 2 expression in (12).…”
Section: Notationmentioning
confidence: 99%