2016
DOI: 10.1109/tc.2015.2506582
|View full text |Cite
|
Sign up to set email alerts
|

GPUvm: GPU Virtualization at the Hypervisor

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
52
0
2

Year Published

2017
2017
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 50 publications
(54 citation statements)
references
References 27 publications
0
52
0
2
Order By: Relevance
“…GPUvm [Suzuki et al 2014] implements both full and para virtualization in the Xen hypervisor by using a Nouveau driver [X.OrgFoundation 2011] in the guest OS side. To isolate multiple VMs on a GPU in full virtualization, GPUvm partitions both physical GPU memory and the MMIO region into several pieces and assigns each portion to an individual VM.…”
Section: Full Virtualizationmentioning
confidence: 99%
See 2 more Smart Citations
“…GPUvm [Suzuki et al 2014] implements both full and para virtualization in the Xen hypervisor by using a Nouveau driver [X.OrgFoundation 2011] in the guest OS side. To isolate multiple VMs on a GPU in full virtualization, GPUvm partitions both physical GPU memory and the MMIO region into several pieces and assigns each portion to an individual VM.…”
Section: Full Virtualizationmentioning
confidence: 99%
“…GPUvm [Suzuki et al 2014] employs the BAND scheduler of Gdev [Kato et al 2012] and solves a flaw of Credit scheduling. The original BAND scheduler distributes credits to each VM based on the assumption that the total utilization of all vGPUs can reach 100%.…”
Section: Algorithms For Scheduling a Single Gpumentioning
confidence: 99%
See 1 more Smart Citation
“…First, small (e.g., sub-millisecond level) but frequent GPU requests, which are common in a wide range of classical HPC applications and emerging applications in real-time analytics, can burden virtualization stacks by frequent context switching between user and hypervisor spaces. Previous research invokes system or hypervisor calls on every GPU request [6], [7], [8], [9], [10], [11], [12], [13], which causes significant per-request trapping costs for small GPU requests. Second, workloads with high CPU-GPU interactivity can cause synchronization bottlenecks between the CPU and GPU schedulers.…”
Section: Introductionmentioning
confidence: 99%
“…However, co-scheduling in itself fails to achieve good fairness because of interference between the CPU and GPU schedulers. Finally, variable request sizes in mixed workloads, which are challenging to handle on non-preemptive GPUs, are either ignored [6], [10], or addressed with reverse engineering methods which are not supported by many GPUs [8], [9], [12], [14]. None of the existing GPU virtualization solutions takes all these factors into account, thus failing to attain acceptable fairness combined with strong performance isolation.…”
Section: Introductionmentioning
confidence: 99%