Proceedings of the 22nd ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming 2017
DOI: 10.1145/3018743.3018748
|View full text |Cite
|
Sign up to set email alerts
|

EffiSha

Abstract: Modern GPUs are broadly adopted in many multitasking environments, including data centers and smartphones. However, the current support for the scheduling of multiple GPU kernels (from different applications) is limited, forming a major barrier for GPU to meet many practical needs. This work for the first time demonstrates that on existing GPUs, efficient preemptive scheduling of GPU kernels is possible even without special hardware support. Specifically, it presents EffiSha, a pure software framework that ena… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 65 publications
(6 citation statements)
references
References 26 publications
0
6
0
Order By: Relevance
“…It schedules GPU kernels by adjusting the number and size of logical TBs spawned in one launch. EffiSha [29] dispatches logical TBs on the basis of the scheduler's decisions. These approaches use the ends of logical TBs as scheduling points.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…It schedules GPU kernels by adjusting the number and size of logical TBs spawned in one launch. EffiSha [29] dispatches logical TBs on the basis of the scheduler's decisions. These approaches use the ends of logical TBs as scheduling points.…”
Section: Related Workmentioning
confidence: 99%
“…To address these problems, we introduce a thin TB scheduler, inspired by the idea of Elastic kernels [38] and EffiSha [29]. Our TB scheduler, which is a software mechanism running inside the device, puts all the TBs in its queue and only executes the same number of TBs as concurrently runnable TBs on the SMs.…”
Section: Thread Block Controlmentioning
confidence: 99%
See 1 more Smart Citation
“…As a result, the behavior of collectives on GPU intrinsically meets the resource-holding condition of a deadlock situation. ❸ There is no publicly accessible official preemptive scheduling support for GPUs, and the GPU-preemption techniques in literature [6,16,24,25,30,42,51,54,59] are not suitable for collectives (see Sec. 6 for details).…”
Section: Background and Motivation 21 Collectives And Deadlocks In Di...mentioning
confidence: 99%
“…Software methods [6,16,24,25,54,59] can be applied directly on commodity GPUs. Wait-based preemption approaches [6,54,59] modify user kernels to insert scheduling points so that user kernels quit more frequently and expose more scheduling opportunities. Lee et al [24,25] and REEF [16] kill the preempted kernel directly to decrease scheduling delay.…”
Section: Related Workmentioning
confidence: 99%