Proceedings of the 2018 International Conference on Supercomputing 2018
DOI: 10.1145/3205289.3205311
|View full text |Cite
|
Sign up to set email alerts
|

Classification-Driven Search for Effective SM Partitioning in Multitasking GPUs

Abstract: Graphics processing units (GPUs) feature an increasing number of streaming multiprocessors (SMs) with each successive generation. At the same time, GPUs are increasingly widely adopted in cloud services and data centers to accelerate general-purpose workloads. Running multiple applications on a GPU in such environments requires effective multitasking support. Spatial multitasking in which independent applications co-execute on different sets of SMs is a promising solution to share GPU resources. Unfortunately,… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 21 publications
(8 citation statements)
references
References 30 publications
0
8
0
Order By: Relevance
“…The performance of the proposed framework for dynamic optimizations is evaluated with the baseline with evenly partitioned spatial multitasking approach [9], [20], [21] and CD-search [13]. Fig.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…The performance of the proposed framework for dynamic optimizations is evaluated with the baseline with evenly partitioned spatial multitasking approach [9], [20], [21] and CD-search [13]. Fig.…”
Section: Resultsmentioning
confidence: 99%
“…Sharing of resources within the core and rest of the off-chip memoryresources have been explored in [12], [11]. We use spatial multitasking for optimizations and compare its performance with the state-of-the-art technique classification driven search (CD-search) [13].…”
Section: Introductionmentioning
confidence: 99%
“…However, if a kernel completes its execution, we end the current epoch and start a new one with an even SM allocation because behavior can be very different across kernels. To reduce the preemption overhead, we adaptively choose between a draining versus context switching policy [5,26]. If, during the epoch, the number of TBs finished on one SM is larger than the number of TBs that can be concurrently executed on one SM, we follow a draining policy; if not, we adopt the context switching policy.…”
Section: Implementing Hsm-based Sm Allocationmentioning
confidence: 99%
“…Managing SMs among concurrent applications in multitasking GPUs received significant attention recently [3,4,14,15,26,42]. These approaches indirectly infer the performance impact of a particular SM allocation; HSM, in contrast, predicts the performance impact of a particular SM allocation.…”
Section: Related Workmentioning
confidence: 99%
“…Graphics Processing Units (GPUs) are throughputoriented co-processors that are witnessing a rapid increase in the amount of computing resources. To avoid keeping these growing resources underutilized and improve performance, concurrent kernel execution (CKE) has been proposed and showed improved GPU throughput and resource utilization [1], [2], [3].…”
Section: Introductionmentioning
confidence: 99%