2021
DOI: 10.48550/arxiv.2105.04026
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

The Modern Mathematics of Deep Learning

Abstract: We describe the new field of mathematical analysis of deep learning. This field emerged around a list of research questions that were not answered within the classical framework of learning theory. These questions concern: the outstanding generalization power of overparametrized neural networks, the role of depth in deep architectures, the apparent absence of the curse of dimensionality, the surprisingly successful optimization performance despite the non-convexity of the problem, understanding what features a… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
22
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 20 publications
(28 citation statements)
references
References 70 publications
(85 reference statements)
0
22
0
Order By: Relevance
“…Despite the tremendous success of deep learning in various applications, also several downsides can be observed. Particular problematic for their applicability are, for instance, instabilities due to adversarial examples in image classification [64] and image reconstruction [4], the fact that deep neural networks still act as a black box lacking explainability [40], [71], and in general a vast gap between theory and practise (see, for instance, [1], [10]). This led to the acknowledgement that for reliable deep learning a better understanding as well as a mathematical theory of deep learning is in great demand.…”
Section: Contributionsmentioning
confidence: 99%
See 1 more Smart Citation
“…Despite the tremendous success of deep learning in various applications, also several downsides can be observed. Particular problematic for their applicability are, for instance, instabilities due to adversarial examples in image classification [64] and image reconstruction [4], the fact that deep neural networks still act as a black box lacking explainability [40], [71], and in general a vast gap between theory and practise (see, for instance, [1], [10]). This led to the acknowledgement that for reliable deep learning a better understanding as well as a mathematical theory of deep learning is in great demand.…”
Section: Contributionsmentioning
confidence: 99%
“…In this section, we give a short introduction to deep learning with a particular focus on solving inverse problems [2], [37], [47], [50], [58]. For a comprehensive depiction of deep learning theory we refer to [30] and [10].…”
Section: Deep Learning For Inverse Problemsmentioning
confidence: 99%
“…Obstacles in the theoretical foundation include the higher-order nonlinear structures due to the stacking of multiple layers and the excessive number of network parameters in state of the art networks. For some recent surveys, see [5,6].…”
Section: 𝑅( A) − 𝑅 𝑛 ( A)mentioning
confidence: 99%
“…, (5,6,7). In the second level, the first and third nodes fire if one or more dependent nodes fire, and the second (dominant) node fires if two or more dependent nodes fire.…”
Section: Small-batch Trainingmentioning
confidence: 99%
“…Deep learning is recognized as a monumentally successful approach to many data-extensive applications in image recognition, natural language processing, and board game programs [1,2,3]. Despite extensive efforts [4,5,6], however, our theoretical understanding of how this increasingly popular machinery works and why it is so effective remains incomplete. This is exemplified by the substantial vacuum between the highly sophisticated training paradigm of modern neural networks and the capabilities of existing theories.…”
Section: Introductionmentioning
confidence: 99%