2021
DOI: 10.48550/arxiv.2111.05193
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Survey on Green Deep Learning

Abstract: In recent years, larger and deeper models are springing up and continuously pushing state-of-the-art (SOTA) results across various fields like natural language processing (NLP) and computer vision (CV). However, despite promising results, it needs to be noted that the computations required by SOTA models have been increased at an exponential rate. Massive computations not only have a surprisingly large carbon footprint but also have negative effects on research inclusiveness and deployment on real-world applic… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
21
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3
1

Relationship

2
7

Authors

Journals

citations
Cited by 18 publications
(21 citation statements)
references
References 126 publications
0
21
0
Order By: Relevance
“…Green AI. Witnessing the exponential growth of computations of big AI models [13,5,42], the concept of Green AI attains mounting attention in recent years [55,67]. Rather than being merely obsessed with accuracy, Green AI advocates making efficiency an important measure of AI models, championing the greener approaches that are more inclusive to the research community.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Green AI. Witnessing the exponential growth of computations of big AI models [13,5,42], the concept of Green AI attains mounting attention in recent years [55,67]. Rather than being merely obsessed with accuracy, Green AI advocates making efficiency an important measure of AI models, championing the greener approaches that are more inclusive to the research community.…”
Section: Related Workmentioning
confidence: 99%
“…To this end, we strive to devise a new and green approach for MIM with hierarchical models, in the spirit of Green AI [55,67]. Our work focuses on extending the asymmetric encoder-decoder architecture of MAE to hierarchical vision transformers, particularly the representative model Swin Transformer [43], for the sake of efficient pre-training on visible patches only.…”
Section: Introductionmentioning
confidence: 99%
“…Consequently, only using model size to assess model efficiency may be inadequate; 3) Actual Inference Time, which is the most intuitive metric for efficiency evaluation. However, since the actual inference time is heavily related to both hardware environment and software implementation, and some algorithms may be hardware-specialized, it is challenging to make a fair comparison between models run on different infrastructures [1233]. In these cases, it is critical to propose new metrics which could comprehensively and faithfully assess model efficiency.…”
Section: Inadequate Metricmentioning
confidence: 99%
“…Using larger models will also likely leads to some improvements. However, we do not run experiments with huge pre-training data and giant models due to environmental considerations [61,62] and we try to make our experiments as "green" as possible.…”
Section: A1 Limitations and Potential Socail Impactsmentioning
confidence: 99%