2019
DOI: 10.3390/math7100992
|View full text |Cite
|
Sign up to set email alerts
|

Universal Function Approximation by Deep Neural Nets with Bounded Width and ReLU Activations

Abstract: This article concerns the expressive power of depth in neural nets with ReLU activations and bounded width. We are particularly interested in the following questions: what is the minimal width wmin(d) so that ReLU nets of width wmin(d) (and arbitrary depth) can approximate any continuous function on the unit cube [0, 1] d aribitrarily well? For ReLU nets near this minimal width, what can one say about the depth necessary to approximate a given function? We obtain an essentially complete answer to these questio… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
157
0
2

Year Published

2019
2019
2023
2023

Publication Types

Select...
8
1
1

Relationship

0
10

Authors

Journals

citations
Cited by 246 publications
(182 citation statements)
references
References 18 publications
2
157
0
2
Order By: Relevance
“…Great progress of deep learning is built on deepening neural networks with structures. Deep nets with different structures have been proved to be universal, i.e., [53], [54] for deep convolutional nets, [14] for deep nets with tree structures and [10] for deep fully-connected neural networks.…”
Section: A Deep Nets With Fixed Structuresmentioning
confidence: 99%
“…Great progress of deep learning is built on deepening neural networks with structures. Deep nets with different structures have been proved to be universal, i.e., [53], [54] for deep convolutional nets, [14] for deep nets with tree structures and [10] for deep fully-connected neural networks.…”
Section: A Deep Nets With Fixed Structuresmentioning
confidence: 99%
“…Different lines of research try to understand the mechanism of deep neural networks from different aspects. For example, a series of work tries to understand how the expressive power of deep neural networks are related to their architecture, including the width of each layer and depth of the network (Telgarsky, 2015(Telgarsky, , 2016Lu et al, 2017;Liang and Srikant, 2016;Yarotsky, 2017Yarotsky, , 2018Hanin, 2017;Hanin and Sellke, 2017). These work shows that multi-layer networks with wide layers can approximate arbitrary continuous function.…”
Section: Introductionmentioning
confidence: 99%
“…That approximation is the partition of the input space into samples that minimizes the error function between the output of the ANN given its training inputs and the training outputs. This is stated mathematically by the universal approximation theorem which implies that any functional mapping between input vectors and output vectors can be approximated to with arbitrary accuracy with an ANN provided that it has a sufficient number of neurons in a sufficient number of layers with a specific activation function [10] [11] [12] [13].…”
Section: A What Is An Artificial Neural Network?mentioning
confidence: 99%