2020
DOI: 10.1016/j.micpro.2020.102991
|View full text |Cite
|
Sign up to set email alerts
|

CoNNa–Hardware accelerator for compressed convolutional neural networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
9

Relationship

0
9

Authors

Journals

citations
Cited by 14 publications
(9 citation statements)
references
References 17 publications
0
4
0
Order By: Relevance
“…Zero-skipping techniques [34] can avoid the multiplication by zero resulting from pruning. Skipping zeros from weights and activations may result in reduced performance efficiency [109]. The large on-chip memory is required in zero-skipping techniques to exploit the parallel processing in hardware acceleration.…”
Section: (C) Structured Block Pruning (Pruning Based On the Lowest Av...mentioning
confidence: 99%
“…Zero-skipping techniques [34] can avoid the multiplication by zero resulting from pruning. Skipping zeros from weights and activations may result in reduced performance efficiency [109]. The large on-chip memory is required in zero-skipping techniques to exploit the parallel processing in hardware acceleration.…”
Section: (C) Structured Block Pruning (Pruning Based On the Lowest Av...mentioning
confidence: 99%
“…The CNN accelerator will speed up both the backbone CNN network and the additional convolutional layers in the SSD Head. This paper used a modified version of the CoNNa CNN HW accelerator, proposed in [22] for this purpose. In contrast, the Puppis HW accelerator will accelerate the remaining calculating functions from the SSD Head: softmax, bounding box, non-maximum suppression, and top-K sorting.…”
Section: System For Hw Acceleration Of Complete Ssd Architecturementioning
confidence: 99%
“…Some authors [140] overcame the on-chip memory limit by considering that the matrix was stored in a dense format, requiring that all weights, including zeros, be loaded. In [141], an architecture was proposed that could skip zeros in both weights and activations. However, the solution had reduced performance efficiency.…”
Section: Hardware-oriented Deep Neural Network Optimizationsmentioning
confidence: 99%