2019
DOI: 10.3390/electronics9010028
|View full text |Cite
|
Sign up to set email alerts
|

Hardware Resource Analysis in Distributed Training with Edge Devices

Abstract: When training a deep learning model with distributed training, the hardware resource utilization of each device depends on the model structure and the number of devices used for training. Distributed training has recently been applied to edge computing. Since edge devices have hardware resource limitations such as memory, there is a need for training methods that use hardware resources efficiently. Previous research focused on reducing training time by optimizing the synchronization process between edge device… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
3
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 23 publications
(24 reference statements)
0
3
0
Order By: Relevance
“…MXNet uses KVStore 1 to synchronize parameters shared among participants during the learning process. To monitor the utilization of pervasive resources, Ganglia [55] is designed to identify memory, CPU, and network requirements of the training and track the hardware usage for each participant. As for the inference phase, authors in [56] designed a hardware prototype targeting distributed deep learning for on-device prediction.…”
Section: Pervasive Framework For Aimentioning
confidence: 99%
“…MXNet uses KVStore 1 to synchronize parameters shared among participants during the learning process. To monitor the utilization of pervasive resources, Ganglia [55] is designed to identify memory, CPU, and network requirements of the training and track the hardware usage for each participant. As for the inference phase, authors in [56] designed a hardware prototype targeting distributed deep learning for on-device prediction.…”
Section: Pervasive Framework For Aimentioning
confidence: 99%
“…The training dataset is divided into mini-lots, and all mini-lots are traversed by tuning the terms in each mini-lot. A tour of all mini-lots corresponds to an epoch [40,41].…”
Section: Descending Gradient With Mini-lotsmentioning
confidence: 99%
“…To evaluate the performance of the proposed model, we and an output layer that predicts values via the fully connected layer. Besides, LeNet-5 works well with handwritten datasets [37], it also reduces the number of parameters and can automatically learn features from raw pixels [38].…”
mentioning
confidence: 99%