2015 IEEE/ACM International Symposium on Code Generation and Optimization (CGO) 2015
DOI: 10.1109/cgo.2015.7054182
|View full text |Cite
|
Sign up to set email alerts
|

Improving GPGPU energy-efficiency through concurrent kernel execution and DVFS

Abstract: Current generation GPUs can accelerate high-performance, computeintensive applications by exploiting massive thread-level parallelism. The high performance, however, comes at the cost of increased power consumption. Recently, commercial GPGPU architectures have introduced support for concurrent kernel execution to better utilize the computational/memory resources and thereby improve overall throughput. In this paper, we argue and experimentally validate the benefits of concurrent kernels towards energyefficien… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
22
0

Year Published

2016
2016
2022
2022

Publication Types

Select...
3
3
3

Relationship

0
9

Authors

Journals

citations
Cited by 40 publications
(22 citation statements)
references
References 10 publications
0
22
0
Order By: Relevance
“…Jiao et al studied the GPU core and memory frequency scaling for two concurrent kernels on the Kepler GT640 GPU [30]. They took a set of kernels from the CUDA SDK and Rodinia benchmark and measured their energy efficiency (GFlops/Watt) with different core-memory frequency settings.…”
Section: A Experimental Studiesmentioning
confidence: 99%
“…Jiao et al studied the GPU core and memory frequency scaling for two concurrent kernels on the Kepler GT640 GPU [30]. They took a set of kernels from the CUDA SDK and Rodinia benchmark and measured their energy efficiency (GFlops/Watt) with different core-memory frequency settings.…”
Section: A Experimental Studiesmentioning
confidence: 99%
“…These generally describe power or energy as linear systems where the cost of executing various types of instructions, accessing different cache hierarchies, disks or network interfaces is found using different methodologies. While some authors have used neural networks to estimate these costs [7,10] the vast majority use multivariable, linear regression. The typical way of describing for example the power usage of GPUs [5,9,18] and CPUs [14,19] is of the form:…”
Section: Background and Related Workmentioning
confidence: 99%
“…From higher to lower, we may distinguish the following. Software: For example, changing the frequency of the GPU core and video memory according to compute‐ and memory‐bound CUDA kernels or combining DVFS with a concurrent kernel execution to improve the performance‐per‐watt behavior compared with their sequential execution Compiler: Wu et al integrated a prototype of a DVFS mechanism into a dynamic compilation system, which is fine‐grained and code‐aware.…”
Section: Introductionmentioning
confidence: 99%