IECON 2020 the 46th Annual Conference of the IEEE Industrial Electronics Society 2020
DOI: 10.1109/iecon43393.2020.9255055
|View full text |Cite
|
Sign up to set email alerts
|

Performance Evaluation of State-of-the-Art Edge Computing Devices for DNN Inference

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(6 citation statements)
references
References 17 publications
0
5
0
Order By: Relevance
“…The deep NN is trained off line; then, through pruning and quantization methods, the groups of artificial neurons that rarely or never fire are removed and the numerical precision of the weights is reduced, so that a reduced model size and a faster computation is achieved at the cost of minimal reduction in prediction accuracy [39]. Based on the parallel characteristics inherent to such algorithms, an FPGA-based or GPUbased implementation is thus highly recommended [40].…”
Section: Next Generation Of Smart Controllers For Electrical Energy Systemsmentioning
confidence: 99%
“…The deep NN is trained off line; then, through pruning and quantization methods, the groups of artificial neurons that rarely or never fire are removed and the numerical precision of the weights is reduced, so that a reduced model size and a faster computation is achieved at the cost of minimal reduction in prediction accuracy [39]. Based on the parallel characteristics inherent to such algorithms, an FPGA-based or GPUbased implementation is thus highly recommended [40].…”
Section: Next Generation Of Smart Controllers For Electrical Energy Systemsmentioning
confidence: 99%
“…The inference time of a DPU on a Zedboard against a CPU, two GPUs and two TPU solutions is compared in [8]. Four common CNNs are tested: Mobilenet v1, Mobilenet v2, Inception v1 and Inception v3.…”
Section: Related Workmentioning
confidence: 99%
“…In the current version of the compiler, all operations after the first non supported operation are also mapped on the CPU [45]. This CPU is slower than the TPU itself, which means that the architecture of the model can have a big impact on the performance of the model on the TPU [8].…”
Section: Tool Flows For Tpumentioning
confidence: 99%
See 2 more Smart Citations