2020
DOI: 10.1007/s10586-020-03144-9
|View full text |Cite
|
Sign up to set email alerts
|

Towards an optimized distributed deep learning framework for a heterogeneous multi-GPU cluster

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
7
1

Relationship

1
7

Authors

Journals

citations
Cited by 16 publications
(8 citation statements)
references
References 6 publications
0
8
0
Order By: Relevance
“…In the method, two parallelism schemas were utilized: data parallelism and model parallelism. In the same direction, Kim et al [88] proposed a distributed DL method based on heterogenous systems. The schema was built by using multiple heterogenous GPUs that worked together.…”
Section: A Parallel Deep Learning On Regular Domainsmentioning
confidence: 99%
“…In the method, two parallelism schemas were utilized: data parallelism and model parallelism. In the same direction, Kim et al [88] proposed a distributed DL method based on heterogenous systems. The schema was built by using multiple heterogenous GPUs that worked together.…”
Section: A Parallel Deep Learning On Regular Domainsmentioning
confidence: 99%
“…There exist a few works that specifically evaluate and/or improve the MPI CCPs for DL, for example, taking into account the special characteristics of the messages that are exchanged in this type of applications [3,4,18,23]. In addition, MPI-based software has been developed for distributed DNN training; for example, MVAPICH2-GDR 1 from Ohio State University or oneAPI 2 from Intel.…”
Section: Mpi Collective Communication Primitivesmentioning
confidence: 99%
“…Tencent's Mariana [36] used DP, which gained a 2:67Â speed increment with four GPUs. In recent years, more paralleled deep learning methods have been brought up [11]. In the aspect of algorithms, several algorithms have been brought up to accelerate multi-GPU implementation or make the inference more accurate [1,26] and faster [7,12].…”
Section: Multi-gpu Parallel Computingmentioning
confidence: 99%