2020
DOI: 10.48550/arxiv.2006.00625
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Neural Networks with Small Weights and Depth-Separation Barriers

Abstract: In studying the expressiveness of neural networks, an important question is whether there are functions which can only be approximated by sufficiently deep networks, assuming their size is bounded. However, for constant depths, existing results are limited to depths 2 and 3, and achieving results for higher depths has been an important open question. In this paper, we focus on feedforward ReLU networks, and prove fundamental barriers to proving such results beyond depth 4, by reduction to open problems and nat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
13
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
5

Relationship

3
2

Authors

Journals

citations
Cited by 7 publications
(13 citation statements)
references
References 20 publications
(48 reference statements)
0
13
0
Order By: Relevance
“…In this work we show that deep networks have significantly more memorization power. Quite a few theoretical works in recent years have explored the beneficial effect of depth on increasing the expressiveness of neural networks (e.g., [23,15,33,22,12,28,38,29,10,34,6,36,35]). The benefits of depth in the context of the VC dimension is implied by, e.g., [3].…”
Section: Related Workmentioning
confidence: 99%
“…In this work we show that deep networks have significantly more memorization power. Quite a few theoretical works in recent years have explored the beneficial effect of depth on increasing the expressiveness of neural networks (e.g., [23,15,33,22,12,28,38,29,10,34,6,36,35]). The benefits of depth in the context of the VC dimension is implied by, e.g., [3].…”
Section: Related Workmentioning
confidence: 99%
“…Overwhelming empirical evidence indicates that deeper networks tend to perform better than shallow ones. Quite a few theoretical works in recent years have explored the beneficial effect of depth on increasing the expressiveness of neural networks (e.g., Martens et al [2013], Eldan and Shamir [2016], Telgarsky [2016], Liang and Srikant [2016], Daniely [2017], Safran and Shamir [2017], Yarotsky [2017], Safran et al [2019], Vardi and Shamir [2020], Bresler and Nagaraj [2020]). A main focus is on depth separation, namely, showing that there is a function f : R d → R that can be approximated by a poly(d)-sized network of a given depth, with respect to some input distribution, but cannot be approximated by poly(d)-sized networks of a smaller depth.…”
Section: Introductionmentioning
confidence: 99%
“…Depth separation between depth 2 and 3 is known Shamir, 2016, Daniely, 2017] 1 already for natural functions. A complexity-theoretic barrier to proving separation between two constant depths beyond depth 4 was established in Vardi and Shamir [2020]. A construction shown by Telgarsky [2016] gives separation between poly(d)-sized networks of a constant depth, and poly(d)-sized networks of some nonconstant depth.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations