2019
DOI: 10.1007/s10766-018-00623-w
|View full text |Cite
|
Sign up to set email alerts
|

Improving the Performance of Distributed MXNet with RDMA

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
0
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 12 publications
(3 citation statements)
references
References 15 publications
0
0
0
Order By: Relevance
“…MXNet is an open-source DL framework developed by Apache. It provides a wide range of tools and libraries for building and deploying DL models, including CNNs [174,175]. MXNet supports both high-level APIs, such as Gluon, and low-level APIs, which allow for greater control over the model architecture.…”
Section: Mxnetmentioning
confidence: 99%
“…MXNet is an open-source DL framework developed by Apache. It provides a wide range of tools and libraries for building and deploying DL models, including CNNs [174,175]. MXNet supports both high-level APIs, such as Gluon, and low-level APIs, which allow for greater control over the model architecture.…”
Section: Mxnetmentioning
confidence: 99%
“…Under certain permissions, researchers may utilize the codes directly or construct new models based on the codes. Theano, fCaffe, gTensorFlow, and MXNet [87] data-intensive operations 140 times faster on GPUs than on CPUs. TensorFlow [88] is an open-source software framework that uses data flow graphs to do numerical computations.…”
Section: Hardware/software Tools Used With Deep Learningmentioning
confidence: 99%
“…Currently, the most common training mode in machine learning is the iterative convergence training mode, the current mainstream distributed implementation process is to iterate gradient descent for each worker, submit the obtained local gradient to the parameter server, enter the synchronization barrier until all workers complete the iteration, and then release the synchronization barrier for the next iteration [4]. As shown in Figure 1, this parameter communication strategy that adds a synchronization barrier to ensure global consistency when updating parameters is called the overall synchronization parallel strategy.…”
Section: Overall Synchronization Parallel Strategymentioning
confidence: 99%