2023
DOI: 10.21203/rs.3.rs-3596530/v1
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Pipeline Parallelism with Reduced Network Communications for Efficient Compute-intensive Neural Network Training

Chanhee Yu,
Kyongseok Park

Abstract: Pipeline parallelism is a distributed deep neural network training method suitable for tasks that consume large amounts of memory. However, this method entails a large amount of overhead because of the dependency between devices in performing forward and backward steps through multiple devices. A method to remove forward step dependency through the all-to-all approach has been proposed for the compute-intensive models; however, this method incurs large overhead when training with a large number of devices and … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 18 publications
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?