Proceedings of the 2019 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays 2019
DOI: 10.1145/3289602.3293945
|View full text |Cite
|
Sign up to set email alerts
|

A Reconfigurable Accelerator for Sparse Convolutional Neural Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(6 citation statements)
references
References 11 publications
0
6
0
Order By: Relevance
“…According to Table V, our implementation achieves 223.4 GOP/s effective performance on sparse Alexnet which shows 2.4X speedup compared with [43] 5 . [49] shows similar performance to our design, but it applies low bit precision which requires less resources. The performance on VGG network is 309.0 GOP/s which is 3.6X-4.8X higher than [16,21].…”
Section: B Performance Analysismentioning
confidence: 62%
See 1 more Smart Citation
“…According to Table V, our implementation achieves 223.4 GOP/s effective performance on sparse Alexnet which shows 2.4X speedup compared with [43] 5 . [49] shows similar performance to our design, but it applies low bit precision which requires less resources. The performance on VGG network is 309.0 GOP/s which is 3.6X-4.8X higher than [16,21].…”
Section: B Performance Analysismentioning
confidence: 62%
“…We also compare our design with previous FPGA accelerators in Table V. [16,21] are dense CNN accelerators and [43,48,49] are sparse CNN accelerators. The performance in Table V represents the effective performance.…”
Section: B Performance Analysismentioning
confidence: 99%
“…As CNN accelerators ( [9], [10], [35] ) are very popular ML accelerators these days, we compare our design with some state-of-the-art CNN accelerators [32]- [34], [36]. Table 3 shows the resources usage of the NBC accelerator is very limited.…”
Section: Methodsmentioning
confidence: 99%
“…It does, however, require optimized implementations of sparse tensor operations [23] to translate the memory and computational savings to practical performance gains on commodity hardware. Nonetheless, many hardware accelerators for DNNs have been proposed with support for unstructured sparsity as their key design goal [24,25,26,27,28,29,30,31,32].…”
Section: Introductionmentioning
confidence: 99%