2020 IEEE High Performance Extreme Computing Conference (HPEC) 2020
DOI: 10.1109/hpec43674.2020.9286209
|View full text |Cite
|
Sign up to set email alerts
|

Architectural Analysis of Deep Learning on Edge Accelerators

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 11 publications
(6 citation statements)
references
References 15 publications
0
6
0
Order By: Relevance
“…The integrated circuits are application-specific (ASICs) that are used to increase specific machine learning tasks by putting processing elements-small digital signal processors (DSPs) with inbuilt memory on a framework and allowing them to communicate and transport data between them. The study in [91] analyzed low-power computer topology built into ML-specific hardware in the context of Chinese handwriting recognition. The work used NVIDIA Jetson AGX Xavier (AGX), Intel Neural Compute Stick 2 (NCS2), and Google Edge TPU architectures have been tested for performance.…”
Section: Tpu-based Edge Hardware Systems and Devicesmentioning
confidence: 99%
“…The integrated circuits are application-specific (ASICs) that are used to increase specific machine learning tasks by putting processing elements-small digital signal processors (DSPs) with inbuilt memory on a framework and allowing them to communicate and transport data between them. The study in [91] analyzed low-power computer topology built into ML-specific hardware in the context of Chinese handwriting recognition. The work used NVIDIA Jetson AGX Xavier (AGX), Intel Neural Compute Stick 2 (NCS2), and Google Edge TPU architectures have been tested for performance.…”
Section: Tpu-based Edge Hardware Systems and Devicesmentioning
confidence: 99%
“…Bianco et al have benchmarked accuracy and inference time of a variety of DNNs on Nvidia Jetson TX1 [27], and found some of the DNNs bottlenecked by the amount of memory available on the device (such as ResNeXt). Kljucaric et al characterized the performance of AlexNet and GoogleNet on several devices: Nvidia jetson AGX Xavier, Intel Neural Compute Stick, and Google Edge TPU [28]. Their results showed best latency for AlexNet is achieved by AGX, while for GoogleNet, TPU is faster.…”
Section: G Architecture-algorithm Insightsmentioning
confidence: 99%
“…As a key result, the paper reports a similar inference performance of Edge TPU compared to the i9-9900k CPU, but with, not surprisingly, a significantly lower power consumption. Kljucaric et al [8] compared the performance and efficiency of NVIDIA Xavier, Edge TPU, and NCS2 for optical character recognition using AlexNet and GoogleNet. The authors reported while NCS2 is more efficient for AlexNet, Edge TPU outperforms with GoogleNet.…”
Section: Related Workmentioning
confidence: 99%