2019
DOI: 10.1109/tit.2019.2927252
|View full text |Cite
|
Sign up to set email alerts
|

A Framework for the Construction of Upper Bounds on the Number of Affine Linear Regions of ReLU Feed-Forward Neural Networks

Abstract: In this work we present a new framework to derive upper bounds on the number regions of feed-forward neural nets with ReLU activation functions. We derive all existing such bounds as special cases, however in a different representation in terms of matrices. This provides new insight and allows a more detailed analysis of the corresponding bounds. In particular, we provide a Jordan-like decomposition for the involved matrices and present new tighter results for an asymptotic setting.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
33
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 15 publications
(33 citation statements)
references
References 2 publications
(1 reference statement)
0
33
0
Order By: Relevance
“…Affine linear regions can be defined as the connected components of R N0 \ H, where H is the set of non-differentiability of the realization 20 Φ (N, R ) (•, θ). A refined analysis on the number of such regions was, for example, conducted by [HvdG19]. It is found that deep ReLU neural networks can exhibit significantly more regions than their shallow counterparts.…”
Section: Alternative Notions Of Expressivitymentioning
confidence: 99%
“…Affine linear regions can be defined as the connected components of R N0 \ H, where H is the set of non-differentiability of the realization 20 Φ (N, R ) (•, θ). A refined analysis on the number of such regions was, for example, conducted by [HvdG19]. It is found that deep ReLU neural networks can exhibit significantly more regions than their shallow counterparts.…”
Section: Alternative Notions Of Expressivitymentioning
confidence: 99%
“…Works in this direction include [45,46,11,47,52,32,36] and earlier works for Boolean circuits and sum-product networks [20,21,10]. The number of linear regions of the functions represented by networks with piecewise linear activations has sparked substantial interest in the study of neural networks, with works including [34,47,33,5,40,23]. Recent works have explored approaches based on tropical geometry [55,9,3] and power diagram subdivisions [6], while others have studied the expectated number of linear regions for typical choices of the parameters in the case of ReLU networks [18,19], empirical enumeration [39], and the relations between linear regions and the behavior of algorithms that are used to select the parameters of neural networks based on data, such as speed of convergence and implicit biases of gradient descent [44,56,26].…”
Section: Introductionmentioning
confidence: 99%
“…These are helpful for the improvement of network structure in this paper. Thanks to the nonlinear characteristics of activation function, neural network with improved activation function has shown good results [18][19][20][21][22][23][24]. In order to learn the distribution characteristics of the nonlinear data better, some improvements of the network are mainly focused on the network's depth.…”
Section: Introductionmentioning
confidence: 99%