2022
DOI: 10.48550/arxiv.2204.01368
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Training Fully Connected Neural Networks is $\exists\mathbb{R}$-Complete

Abstract: We consider the algorithmic problem of finding the optimal weights and biases for a twolayer fully connected neural network to fit a given set of data points. This problem is known as empirical risk minimization in the machine learning community. We show that the problem is ∃R-complete. This complexity class can be defined as the set of algorithmic problems that are polynomial-time equivalent to finding real roots of a polynomial with integer coefficients. Our results hold even if the following restrictions ar… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2
2
1

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 48 publications
0
3
0
Order By: Relevance
“…Important ∃R-completeness results include the realizability of abstract order types [48,59] and geometric linkages [53], as well as the recognition of geometric segment [36,44], unit-disk [34,46], and ray intersection graphs [19]. More results appeared in the graph drawing community [22,23,41,54], regarding the Hausdorff distance [33], regarding polytopes [21,51], the study of Nash-equilibria [6,9,10,25,55], training neural networks [3,8], matrix factorization [20,56,57,58,61], or continuous constraint satisfaction problems [47]. In computational geometry, we would like to mention geometric packing [4], the art gallery problem [2], and covering polygons with convex polygons [1].…”
Section: Background and Related Workmentioning
confidence: 99%
“…Important ∃R-completeness results include the realizability of abstract order types [48,59] and geometric linkages [53], as well as the recognition of geometric segment [36,44], unit-disk [34,46], and ray intersection graphs [19]. More results appeared in the graph drawing community [22,23,41,54], regarding the Hausdorff distance [33], regarding polytopes [21,51], the study of Nash-equilibria [6,9,10,25,55], training neural networks [3,8], matrix factorization [20,56,57,58,61], or continuous constraint satisfaction problems [47]. In computational geometry, we would like to mention geometric packing [4], the art gallery problem [2], and covering polygons with convex polygons [1].…”
Section: Background and Related Workmentioning
confidence: 99%
“…However, for a full theoretical understanding of this fundamental machine learning model it is necessary to understand what functions can be exactly expressed with different NN architectures. For instance, insights about exact representability have boosted our understanding of the computational complexity of the task to train an NN with respect to both, algorithms [4,36] and hardness results [9,18,20]. It is known that a function can be expressed with a ReLU NN if and only if it is continuous and piecewise linear (CPWL) [4].…”
Section: Introductionmentioning
confidence: 99%
“…However, the implications on the computational complexity are limited since their result requires the number of hidden neurons to be very large. Bertschinger, Hertrich, Jungeblut, Miltzow, and Weber (2022) show that training 2-layer neural networks is complete for the complexity class ∃R (existential theory of the reals), implying that the problem is presumably not contained in NP. They generalize a previous result by Abrahamsen, Kleist, and Miltzow (2021), who showed the same fact for specifically designed, more complex architectures.…”
Section: Introductionmentioning
confidence: 99%