2014
DOI: 10.1145/2644865.2541963
|View full text |Cite
|
Sign up to set email alerts
|

Disengaged scheduling for fair, protected access to fast computational accelerators

Abstract: Today's operating systems treat GPUs and other computational accelerators as if they were simple devices, with bounded and predictable response times. With accelerators assuming an increasing share of the workload on modern machines, this strategy is already problematic, and likely to become untenable soon. If the operating system is to enforce fair sharing of the machine, it must assume responsibility for accelerator scheduling and resource management. Fair, safe scheduling is a particular challenge… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(10 citation statements)
references
References 16 publications
0
10
0
Order By: Relevance
“…After launching the GPU kernel, the host thread continues its execution 3 . Right before it reaches the next waiting point, it gets into a while loop, in which, it repeatedly checks the KState until it becomes "DONE".…”
Section: Basic Implementation Of the Runtimementioning
confidence: 99%
See 2 more Smart Citations
“…After launching the GPU kernel, the host thread continues its execution 3 . Right before it reaches the next waiting point, it gets into a while loop, in which, it repeatedly checks the KState until it becomes "DONE".…”
Section: Basic Implementation Of the Runtimementioning
confidence: 99%
“…Moreover, the default scheduling is oblivious to the priorities of kernels. Numerous studies [1][2][3][4] have shown that the problematic way to manage GPU causes serious unfairness, response delays, and low GPU utilizations.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Table 1 shows the statistics generated by the agents. Rinnegan does not yet handle directly accessible accelerators, but disengaged schedulers [41] can be extended to implement the agent functionality for such devices. GPU agent.…”
Section: Accelerator Agentsmentioning
confidence: 99%
“…It may be up to the device driver for an accelerator to make scheduling decisions, which may not be coordinated with CPU scheduling. While there have been research systems such as PTask and others [25,41,56] that perform scheduling for tasks on a single processing unit, they are unable to select between multiple possible units for a task. Conversely, application runtimes for heterogeneous systems can run a task on different processing units [6,39], but support only static heterogeneity, where the performance of a processing unit does not vary over time.…”
Section: Introductionmentioning
confidence: 99%