2017
DOI: 10.1016/j.neunet.2017.07.002
|View full text |Cite
|
Sign up to set email alerts
|

Error bounds for approximations with deep ReLU networks

Abstract: We study expressive power of shallow and deep neural networks with piece-wise linear activation functions. We establish new rigorous upper and lower bounds for the network complexity in the setting of approximations in Sobolev spaces. In particular, we prove that deep ReLU networks more efficiently approximate smooth functions than shallow networks. In the case of approximations of 1D Lipschitz functions we describe adaptive depth-6 network architectures more efficient than the standard shallow architecture.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

22
905
2
2

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 937 publications
(986 citation statements)
references
References 16 publications
22
905
2
2
Order By: Relevance
“…Similar results for approximating functions in W k,p ([−1, 1] d ) with p < ∞ using ReLU DNNs are given by Petersen and Voigtlaender[13]. The significance of the works by Yarotsky [12] and Peterson and Voigtlaender [13] is that by using a very simple rectified nonlinearity, DNNs can obtain high order approximation property. Shallow networks do not hold such a good property.…”
supporting
confidence: 64%
“…Similar results for approximating functions in W k,p ([−1, 1] d ) with p < ∞ using ReLU DNNs are given by Petersen and Voigtlaender[13]. The significance of the works by Yarotsky [12] and Peterson and Voigtlaender [13] is that by using a very simple rectified nonlinearity, DNNs can obtain high order approximation property. Shallow networks do not hold such a good property.…”
supporting
confidence: 64%
“…If σfalse(xfalse)=maxfalse(x,0false), the multilayer feedforward neural network is the deep ReLU network. Yarotsky shows that deep ReLU networks can implement multiplication in the following proposition …”
Section: Notation and Preliminary Resultsmentioning
confidence: 99%
“…With the successful applications of deep learning in computer vision, speech recognition, natural language processing, and other fields, the expressive ability of deep neural networks as the theoretical foundation of deep learning has attracted more and more attention. [1][2][3][4][5][6][7] The ability to overcome the "curse of dimensionality" (the number of parameters needed to support the result grows exponentially with the dimensionality) is considered one of the advantages of deep neural networks. [8][9][10][11][12][13][14][15] Recently, Lee H et al 12…”
Section: Introductionmentioning
confidence: 99%
“…Deep ReLU with LSTM cells have became popular architectures as they can capture long‐range dependencies and nonlinearities. Their popularity stems from the fact that they can efficiently approximate highly multivariate functions with small number of neurons at each layer …”
Section: Deep Learningmentioning
confidence: 99%
“…Their popularity stems from the fact that they can efficiently approximate highly multivariate functions with small number of neurons at each layer. [32][33][34]…”
Section: Deep Learningmentioning
confidence: 99%