2021 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW) 2021
DOI: 10.1109/ipdpsw52791.2021.00110
|View full text |Cite
|
Sign up to set email alerts
|

A Flexible Research-Oriented Framework for Distributed Training of Deep Neural Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
1

Relationship

3
2

Authors

Journals

citations
Cited by 6 publications
(7 citation statements)
references
References 13 publications
0
7
0
Order By: Relevance
“…A second article on PyDTNN [3] provided practical evidence that the distributed training on GPUs using PyDTNN attains similar accuracy and parallel performance to those achieved by Tensor-Flow+Horovod on GPUs. In that case, the GPU backend of PyDTNN was used, which internally calls the NVIDIA cuDNN library to perform the model layers related operations.…”
Section: Comparison With Tensorflow+horovodmentioning
confidence: 99%
See 4 more Smart Citations
“…A second article on PyDTNN [3] provided practical evidence that the distributed training on GPUs using PyDTNN attains similar accuracy and parallel performance to those achieved by Tensor-Flow+Horovod on GPUs. In that case, the GPU backend of PyDTNN was used, which internally calls the NVIDIA cuDNN library to perform the model layers related operations.…”
Section: Comparison With Tensorflow+horovodmentioning
confidence: 99%
“…The use of the "reshape" operator A ≡ Reshape(F ) there re-arranges the input 4D filter tensor F as the 2D matrix A. In addition, the reshape followed by a transpose O ≡ Reshape(C ) T (1,2,0,3) , where the superindex (1, 2, 0, 3) specifies the permutation applied to the dimensions of Reshape(C ), re-organizes the resulting C matrix back into the 4D output tensor O .…”
Section: Convolution Operators Via Gemm: the Im2col In Fpmentioning
confidence: 99%
See 3 more Smart Citations