21st IEEE Real-Time and Embedded Technology and Applications Symposium 2015
DOI: 10.1109/rtas.2015.7108420
|View full text |Cite
|
Sign up to set email alerts
|

GPES: a preemptive execution system for GPGPU computing

Abstract: Graphics processing units (GPUs) are being widely used as co-processors in many application domains to accelerate general-purpose workloads that are computationally intensive, known as GPGPU computing. Real-time multi-tasking support is a critical requirement for many emerging GPGPU computing domains. However, due to the asynchronous and non-preemptive nature of GPU processing, in multi-tasking environments, tasks with higher priority may be blocked by lower priority tasks for a lengthy duration. This severely… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
19
0

Year Published

2017
2017
2019
2019

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 53 publications
(21 citation statements)
references
References 12 publications
0
19
0
Order By: Relevance
“…This situation can cause unfairness between multiple kernels and significantly deteriorate the system responsiveness. Existing GPU scheduling methods address this issue by either killing a long running kernel [Menychtas et al 2014] or providing a kernel split tool [Basaran and Kang 2012;Zhou et al 2015;Margiolas and O'Boyle 2016]. The Pascal architecture allows GPU kernels to be interrupted at instruction-level granularity by saving and restoring each GPU context to and from the GPU's DRAM.…”
Section: Algorithms For Scheduling a Single Gpumentioning
confidence: 99%
“…This situation can cause unfairness between multiple kernels and significantly deteriorate the system responsiveness. Existing GPU scheduling methods address this issue by either killing a long running kernel [Menychtas et al 2014] or providing a kernel split tool [Basaran and Kang 2012;Zhou et al 2015;Margiolas and O'Boyle 2016]. The Pascal architecture allows GPU kernels to be interrupted at instruction-level granularity by saving and restoring each GPU context to and from the GPU's DRAM.…”
Section: Algorithms For Scheduling a Single Gpumentioning
confidence: 99%
“…The literature shows this technique to be feasible with low performance overhead and significant benefits in terms of responsiveness [36]. We have adopted the implementation details from GPES [37]. When a vGPU has a long running kernel, FairGV divides the kernel into a set of sub-kernels so that a sub-kernel is executed by a specified number of thread blocks.…”
Section: Collaborative Schedulingmentioning
confidence: 99%
“…The scheme has been studied in real-time community, represented by the recent proposals of GPES [7], and PKM [6]. A challenge in this scheme is to select the appropriate slice size.…”
Section: Related Workmentioning
confidence: 99%
“…The second class of work is about granularity. They use kernel slicing [6,7] to break one GPU kernel into many smaller ones. The reduced granularity increases the flexibility in kernel scheduling, and may help shorten the time that a kernel has to wait before it can get launched.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation