2022
DOI: 10.1007/s10915-021-01718-2
|View full text |Cite
|
Sign up to set email alerts
|

Constructive Deep ReLU Neural Network Approximation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
18
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 18 publications
(18 citation statements)
references
References 31 publications
0
18
0
Order By: Relevance
“…Moreover, we also consider the case of analytic functions and will prove that a two hidden layer tanh neural network suffices to approximate an analytic function at an exponential rate, in terms of the network width, even in Sobolev norms. This result provides an improvement over available results for the approximation of analytic functions by ReLU neural networks [65,55,22] and also neural networks with smooth activation functions [44] and further illustrate the powers of rather shallow tanh networks at approximating smooth functions. Finally, we also derive explicit bounds on the width of the tanh neural networks as well as asymptotic bounds on their weights, thus paving the way for bounds on the generalization error for these neural networks.…”
Section: Introductionmentioning
confidence: 57%
See 2 more Smart Citations
“…Moreover, we also consider the case of analytic functions and will prove that a two hidden layer tanh neural network suffices to approximate an analytic function at an exponential rate, in terms of the network width, even in Sobolev norms. This result provides an improvement over available results for the approximation of analytic functions by ReLU neural networks [65,55,22] and also neural networks with smooth activation functions [44] and further illustrate the powers of rather shallow tanh networks at approximating smooth functions. Finally, we also derive explicit bounds on the width of the tanh neural networks as well as asymptotic bounds on their weights, thus paving the way for bounds on the generalization error for these neural networks.…”
Section: Introductionmentioning
confidence: 57%
“…Exponential convergence (in terms of network size) of neural networks for analytic functions in the L ∞ -norm was first proven in [44] for neural networks with smooth activation functions and in [65] for ReLU neural networks. In [55,22], the authors prove exponential convergence in W 1,∞ -norm for ReLU neural networks. We compare our results for approximation of analytic functions with these papers in Table 2.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…In the last several years, there has been a number of interesting papers that addressed the role of depth and architecture of deep neural networks in approximating functions that possess special regularity properties such as analytic functions [20,38], differentiable functions [45,52], oscillatory functions [29], functions in Sobolev or Besov spaces [1,27,30,53]. High-dimensional approximations by deep neural networks have been studied in [39,48,16,17], and their applications to high-dimensional PDEs in [47,21,43,31,25,26,28]. Most of these papers used deep ReLU (Rectified Linear Unit) neural networks since the rectified linear unit is a simple and preferable activation function in many applications.…”
Section: Introductionmentioning
confidence: 99%
“…Recently, a number of works have been devoted to various problems and methods of deep neural network approximation for parametric and stochastic PDEs such as dimensionality reduction [51], deep neural network expression rates for the Taylor generalized polynomial chaos expansion (gpc) of solutions to parametric elliptic PDEs [46], reduced basis methods [36] the problem of learning the discretized parameter-to-solution map in practice [24], Bayesian PDE inversion [42,32,31], etc. In particular, in [46] the authors proved dimension-independent deep neural network expression rate bounds of the uniform approximation of solution to parametric elliptic PDE with affine inputs on I ∞ := [−1, 1] ∞ based on n-term truncations of the non-orthogonal Taylor gpc expansion.…”
Section: Introductionmentioning
confidence: 99%