Proceedings of the 28th ACM International Conference on Architectural Support for Programming Languages and Operating Systems, 2023
DOI: 10.1145/3575693.3575703
|View full text |Cite
|
Sign up to set email alerts
|

Mobius: Fine Tuning Large-Scale Models on Commodity GPU Servers

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(1 citation statement)
references
References 11 publications
0
1
0
Order By: Relevance
“…The method proposed by Wang et al [68] involves decomposing the original communication and computational operations into more fine-grained tasks, thereby achieving an overlap between communication and computation that effectively reduces data communication overhead. Mobius [69] introduces a pipeline strategy for heterogeneous memory, which overlaps communication with computation by prefetching data from the CPU to the GPU memory for the next stage. Additionally, it employs a Cross-mapping strategy to reduce communication contention, further optimizing overall performance.…”
Section: Communication Optimizationmentioning
confidence: 99%
“…The method proposed by Wang et al [68] involves decomposing the original communication and computational operations into more fine-grained tasks, thereby achieving an overlap between communication and computation that effectively reduces data communication overhead. Mobius [69] introduces a pipeline strategy for heterogeneous memory, which overlaps communication with computation by prefetching data from the CPU to the GPU memory for the next stage. Additionally, it employs a Cross-mapping strategy to reduce communication contention, further optimizing overall performance.…”
Section: Communication Optimizationmentioning
confidence: 99%