2021
DOI: 10.1109/mspec.2021.9563954
|View full text |Cite
|
Sign up to set email alerts
|

Deep Learning's Diminishing Returns: The Cost of Improvement is Becoming Unsustainable

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
87
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 93 publications
(106 citation statements)
references
References 0 publications
0
87
0
Order By: Relevance
“…The faulty estimates in [2] are understandable given the lack of access to internal information. It is likewise understandable that those estimates were propagated in other papers, like [1,6,7,17,18,19,20].…”
Section: Related Workmentioning
confidence: 93%
See 2 more Smart Citations
“…The faulty estimates in [2] are understandable given the lack of access to internal information. It is likewise understandable that those estimates were propagated in other papers, like [1,6,7,17,18,19,20].…”
Section: Related Workmentioning
confidence: 93%
“…The opening quote in Section 1 is based on a 2019 study from the University of Massachusetts (UMass) that estimated the environmental impact of training [2]. More than 1000 papers cite this paper as the source for the impact on carbon emissions of ML models, e.g., [1,6,7,17,18,19,20]. The study calculated the energy consumed and carbon footprint of the NAS by [8] that led to Evolved Transformer.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…The main strength of these techniques is that their classification accuracy typically improves as they are trained on more data, scaling to datasets containing billions of images (Mahajan et al, 2018 ). However, this strength is also becoming a main point of critique, as an exponential increase in compute (and energy) resources is required for marginal gains (Thompson et al, 2021 ). Moreover, these classifiers are known to be vulnerable to ambiguous and adversarial samples (Gilmer et al, 2018 ), and are restricted to object categories known and seen during training.…”
Section: Introductionmentioning
confidence: 99%
“…The power consumption of conventional, transistor-based computers is unsustainably high, particularly with the burgeoning of fields such as neuromorphic computing and machine learning (Thompson et al, 2021). For example, the power density of typical supercomputer might be in the order of 100 Wcm −2 .…”
Section: Introductionmentioning
confidence: 99%