2020
DOI: 10.4208/cicp.oa-2020-0149
|View full text |Cite
|
Sign up to set email alerts
|

Deep Network Approximation Characterized by Number of Neurons

Abstract: This paper develops simple feed-forward neural networks that achieve the universal approximation property for all continuous functions with a fixed finite number of neurons. These neural networks are simple because they are designed with a simple and computable continuous activation function σ leveraging a triangularwave function and a softsign function. We prove that σ-activated networks with width 36d(2d + 1) and depth 11 can approximate any continuous function on a d-dimensioanl hypercube within an arbitrar… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
68
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
1
1

Relationship

1
7

Authors

Journals

citations
Cited by 96 publications
(71 citation statements)
references
References 70 publications
0
68
0
Order By: Relevance
“…Based on the proof sketch stated just above, we are ready to give the detailed proof of theorem 2 following similar ideas as in our previous work (Shen et al, 2019;Shen et al, 2020;Lu et al, 2020). The main idea of our proof is to reduce high-dimensional approximation to one-dimensional approximation via a projection.…”
Section: Proof Of Theorem 1 Theorem 1 Is An Immediate Consequence Of Theoremmentioning
confidence: 97%
“…Based on the proof sketch stated just above, we are ready to give the detailed proof of theorem 2 following similar ideas as in our previous work (Shen et al, 2019;Shen et al, 2020;Lu et al, 2020). The main idea of our proof is to reduce high-dimensional approximation to one-dimensional approximation via a projection.…”
Section: Proof Of Theorem 1 Theorem 1 Is An Immediate Consequence Of Theoremmentioning
confidence: 97%
“…17 In the following years, some analysis of the function approximation using single hidden layer networks have been carried out successively; see, e.g., 12,[18][19][20][21][22] However, few studies have focused on the choice of the hyperparameters of the network and how it influences the numerical solution of differential equations. Nevertheless we notice the work by Shen et al, 23 who have studied deep network approximation, characterized by the number of neurons.…”
Section: Introductionmentioning
confidence: 95%
“…The recent breakthrough of deep learning has attracted much research on the approximation theory of deep neural networks. The approximation rates of ReLU deep neural networks are well studied for many function classes, such as continuous functions (Yarotsky, 2017(Yarotsky, , 2018Shen et al, 2020), smooth functions (Yarotsky and Zhevnerchuk, 2020;Lu et al, 2021), piecewise smooth functions (Petersen and Voigtlaender, 2018), shift-invariant spaces and band-limited functions (Montanelli et al, 2019).…”
Section: Introductionmentioning
confidence: 99%
“…The recent works (Neyshabur et al, 2015;Bartlett et al, 2017;Golowich et al, 2018;Barron and Klusowski, 2019) also show that the sample complexity of deep neural networks can be controlled by certain norms of the weights. However, in the approximation theory literature, the approximation rates of deep neural networks are characterized by the number of weights (Yarotsky, 2017(Yarotsky, , 2018Yarotsky and Zhevnerchuk, 2020) or the number of neurons (Shen et al, 2020;Lu et al, 2021), rather than the size of weights.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation