2018
DOI: 10.1016/j.neunet.2018.08.019
|View full text |Cite
|
Sign up to set email alerts
|

Optimal approximation of piecewise smooth functions using deep ReLU neural networks

Abstract: We study the necessary and sufficient complexity of ReLU neural networks - in terms of depth and number of weights - which is required for approximating classifier functions in an L-sense. As a model class, we consider the set E(R) of possibly discontinuous piecewise C functions f:[-12,12]→R, where the different "smooth regions" of f are separated by C hypersurfaces. For given dimension d≥2, regularity β>0, and accuracy ε>0, we construct artificial neural networks with ReLU activation function that approximate… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

9
412
2

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 382 publications
(421 citation statements)
references
References 51 publications
(91 reference statements)
9
412
2
Order By: Relevance
“…Similar results for approximating functions in W k,p ([−1, 1] d ) with p < ∞ using ReLU DNNs are given by Petersen and Voigtlaender[13]. The significance of the works by Yarotsky [12] and Peterson and Voigtlaender [13] is that by using a very simple rectified nonlinearity, DNNs can obtain high order approximation property. Shallow networks do not hold such a good property.…”
supporting
confidence: 64%
See 2 more Smart Citations
“…Similar results for approximating functions in W k,p ([−1, 1] d ) with p < ∞ using ReLU DNNs are given by Petersen and Voigtlaender[13]. The significance of the works by Yarotsky [12] and Peterson and Voigtlaender [13] is that by using a very simple rectified nonlinearity, DNNs can obtain high order approximation property. Shallow networks do not hold such a good property.…”
supporting
confidence: 64%
“…Shallow networks do not hold such a good property. Other works show ReLU DNNs have high-order approximation property include the work by E and Wang [14] and the recent work by Opschoor et al [15], the latter one relates ReLU DNNs to high-order finite element methods.A basic fact used in the error estimate given in [12] and [13] is that x 2 , x y can be approximated by a ReLU network with O (log |ε|) layers. To remove this approximation error and the extra factor O (log |ε|) in the size of neural networks, we proposed to use rectified power units (RePU) to construct exact neural network representations of polynomials [16].…”
mentioning
confidence: 99%
See 1 more Smart Citation
“…Such constructive approaches can further be found in [8], in [14] for β-cartoon-like functions, in [20] for (b, ε)holomorphic maps, and in [15] for high-frequent sinusoidal functions.…”
Section: Settingmentioning
confidence: 99%
“…In this context it becomes relevant to not only show how well a given function of interest can be approximated by neural networks but also to extend the study to the derivative of this function. A number of recent publications [13], [14], [15] have investigated the required size of a network which is sufficient to approximate certain interesting (classes of) functions within a given accuracy. This is achieved, first, by considering the approximation of basic functions by very simple networks and, subsequently, by combining those networks in order to approximate more difficult structures.…”
Section: Introductionmentioning
confidence: 99%