2020
DOI: 10.48550/arxiv.2007.05558
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

The Computational Limits of Deep Learning

Abstract: Deep learning's recent history has been one of achievement: from triumphing over humans in the game of Go to world-leading performance in image recognition, voice recognition, translation, and other tasks. But this progress has come with a voracious appetite for computing power. This article reports on the computational demands of Deep Learning applications in five prominent application areas and shows that progress in all five is strongly reliant on increases in computing power. Extrapolating forward this rel… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
90
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
5

Relationship

0
10

Authors

Journals

citations
Cited by 114 publications
(115 citation statements)
references
References 44 publications
(56 reference statements)
0
90
0
Order By: Relevance
“…Unfortunately, despite all of the recent success, modern hardware still greatly restricts the practicality of certain machine learning models. Machine learning, deep learning in particular, can be very computationally expensive, sometimes requiring hours, days, or even months of training time on today's computers [4]. Moreover, conventional computers are beginning to approach physical limitations that will slow their improvements in years to come [5].…”
Section: Introductionmentioning
confidence: 99%
“…Unfortunately, despite all of the recent success, modern hardware still greatly restricts the practicality of certain machine learning models. Machine learning, deep learning in particular, can be very computationally expensive, sometimes requiring hours, days, or even months of training time on today's computers [4]. Moreover, conventional computers are beginning to approach physical limitations that will slow their improvements in years to come [5].…”
Section: Introductionmentioning
confidence: 99%
“…It is estimated that inference accounts for up to 90% of the costs [Thomas, 2020]. There are several studies about training computation and its environmental impact [Amodei and Hernandez, 2018, Gholami et al, 2021a, Canziani et al, 2017, Li et al, 2016, Anthony et al, 2020, Thompson et al, 2020 but there are very few focused on inference costs and their associated energy consumption.…”
Section: Introductionmentioning
confidence: 99%
“…On the other hand, it is understood that artificial intelligence systems need not mimic the low-level architecture of the brain cells, but rather get inspirations from abstract properties of human intelligence [15]. This becomes especially important when considering that adopting black-box deep neural network architectures results in using overly complex models of a great many parameters in the expense of time, energy, data, memory and computational resources [16], [17], Even in the applications when complexity is not an issue, the lack of interpretability and mathematical understanding, and the vulnerability to small perturbations and adversarial attacks [18]- [20], has led to an emerging hesitation in the use of deep learning models outside common benchmark datasets [21], [22], and, especially, in security critical applications. These models are hard to analyze with current mathematical tools, hard to train with current optimization methods, and their design relies solely in experimental heuristics.…”
Section: Introductionmentioning
confidence: 99%