2022
DOI: 10.1109/tpds.2021.3137867
|View full text |Cite
|
Sign up to set email alerts
|

Dynamic GPU Energy Optimization for Machine Learning Training Workloads

Abstract: GPUs are widely used to accelerate the training of machine learning workloads. As modern machine learning models become increasingly larger, they require a longer time to train, leading to higher GPU energy consumption. This paper presents GPOEO, an online GPU energy optimization framework for machine learning training workloads. GPOEO dynamically determines the optimal energy configuration by employing novel techniques for online measurement, multi-objective prediction modeling, and search optimization. To ch… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 15 publications
(9 citation statements)
references
References 19 publications
0
4
0
Order By: Relevance
“…Graphical Processing Units (GPU) are processors capable of parallel processing instructions. Standard GPU deep learning speedup techniques include convolutional layer reuse, featuremap reuse and filter reuse, and memory access is a common bottleneck [163]. The basic idea is that functions that are computed many times should be optimized on all levels, from high to low, including the instruction set level.…”
Section: Stack Optimizations For Deep Learningmentioning
confidence: 99%
“…Graphical Processing Units (GPU) are processors capable of parallel processing instructions. Standard GPU deep learning speedup techniques include convolutional layer reuse, featuremap reuse and filter reuse, and memory access is a common bottleneck [163]. The basic idea is that functions that are computed many times should be optimized on all levels, from high to low, including the instruction set level.…”
Section: Stack Optimizations For Deep Learningmentioning
confidence: 99%
“…As an example, for V100 average energy savings of 24%–33%, for EDP 23%–27% with performance loss of 13%–21% and for EDS (k=2$$ k=2 $$) 23.5%–27.3% with performance loss of 4.5%–13.8% were observed. In Reference 36, a GPOEO solution was proposed, developed specifically for iterative machine learning applications. The tool measures performance counter as well as energy online and time and energy models are used to find best predicted configuration that optimizes a function of time and energy.…”
Section: Related Workmentioning
confidence: 99%
“…This approach considers both the model's quality and energy consumption. In [21], the authors presented an online GPU energy optimization framework for assessing iterative ML workloads and automatically predicting the best energy configuration.…”
Section: A Cpu-gpu Based Systemsmentioning
confidence: 99%