2017
DOI: 10.48550/arxiv.1712.08688
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

New error bounds for deep networks using sparse grids

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
8
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(8 citation statements)
references
References 0 publications
0
8
0
Order By: Relevance
“…For these networks, approximation bounds for classes of smooth functions have been established in [61] and for piecewise smooth functions in [47]. Connections to sparse grids and the associated approximation rates were established in [43] and connections to linear finite element approximation were reported in [27]. It was also discovered that the depth of neural networks, i.e., the number of layers, crucially influences the approximation capabilities of these networks in the sense that deeper networks are more efficient approximators [13,42,47,50].…”
Section: Introductionmentioning
confidence: 99%
“…For these networks, approximation bounds for classes of smooth functions have been established in [61] and for piecewise smooth functions in [47]. Connections to sparse grids and the associated approximation rates were established in [43] and connections to linear finite element approximation were reported in [27]. It was also discovered that the depth of neural networks, i.e., the number of layers, crucially influences the approximation capabilities of these networks in the sense that deeper networks are more efficient approximators [13,42,47,50].…”
Section: Introductionmentioning
confidence: 99%
“…In addition, we are interested in understanding the SGD method in solving the non-convex optimization problem. The TensorFlow [1] provides an efficient tool to calculate the partial derivatives in (20), which will be used in our implementation.…”
Section: Numerical Examplementioning
confidence: 99%
“…In recent years, deep learning methods have achieved unprecedented successes in various application fields, including computer vision, speech recognition, natural language processing, audio recognition, social network filtering, and bioinformatics, where they have produced results comparable to and in some cases superior to human experts [17,12]. Motivated by these exciting progress, there are increased new research interests in the literature for the application of deep learning methods for scientific computation, including approximating multivariate functions and solving differential equations using the deep neural network; see [13,20,27,28,15,32] and references therein.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…A practical issue of such a convex nonlinear optimization problem is the difficulty in identifying the global minimizer using numerical methods. While this issue is not solved, recent advances in deep learning theory show that the deep neural network (DNN), as a composition of multiple linear transformations and simple nonlinear activation functions, has the capacity of approximating various kinds of functions, overcoming or mitigating the curse of dimensionality [46,14,48,51,67,37,47,23,58]. Besides, it is shown that with over-parametrization and random initialization, the DNN-based least square optimization achieves a global minimizer by gradient descent with a linear convergence rate in both the setting of regression [27,11,68,7,43,40,9,8] and PDE solvers [41,34].…”
Section: Introductionmentioning
confidence: 99%