2017
DOI: 10.1145/3068281
|View full text |Cite
|
Sign up to set email alerts
|

GPU Virtualization and Scheduling Methods

Abstract: The integration of graphics processing units (GPUs) on high-end compute nodes has established a new accelerator-based heterogeneous computing model, which now permeates high performance computing. The same paradigm nevertheless has limited adoption in cloud computing or other large-scale distributed computing paradigms. Heterogeneous computing with GPUs can benefit the Cloud by reducing operational costs and improving resource and energy efficiency. However, such a paradigm shift would require effective method… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
63
0
1

Year Published

2017
2017
2021
2021

Publication Types

Select...
6
2
2

Relationship

2
8

Authors

Journals

citations
Cited by 85 publications
(64 citation statements)
references
References 127 publications
0
63
0
1
Order By: Relevance
“…In fog/edge computing, containers are widely used because they realize lightweight virtualization. However, efficient GPU resource management in containers has not been explored sufficiently, compared to research in virtual machines [100]. In fog/edge devices, GPUs can be used for data analytics and to assist deep learning algorithms.…”
Section: Discussionmentioning
confidence: 99%
“…In fog/edge computing, containers are widely used because they realize lightweight virtualization. However, efficient GPU resource management in containers has not been explored sufficiently, compared to research in virtual machines [100]. In fog/edge devices, GPUs can be used for data analytics and to assist deep learning algorithms.…”
Section: Discussionmentioning
confidence: 99%
“…However, the special case happens when t = δ i because at this time, if task i has not completed, it is dropped. For the purposes of calculating PCT (i, j) using Equation 5, PCT (i − 1, j) is guaranteed to be complete by its deadline. Therefore, as Equation 5 shows, all the impulses after δ i are aggregated into the impulse at t = δ i .…”
Section: Calculating Task Completion Time In the Presence Of Taskmentioning
confidence: 99%
“…A successful middleware to implement this approach is rCUDA [24], which enables the concurrent remote usage of CUDAenabled devices in a transparent way. An extensive survey for GPU virtualization techniques and scheduling methods is provided in [25]. Although there exist several scheduling methods to schedule job tasks into GPUs, varying from priority-based to load-balancing-based approaches, they perform fine-grained scheduling, being implemented at hypervisor or OS level.…”
Section: Related Workmentioning
confidence: 99%