2021 IEEE International Symposium on Circuits and Systems (ISCAS) 2021
DOI: 10.1109/iscas51556.2021.9401214
|View full text |Cite
|
Sign up to set email alerts
|

ChewBaccaNN: A Flexible 223 TOPS/W BNN Accelerator

Abstract: Binary Neural Networks enable smart IoT devices, as they significantly reduce the required memory footprint and computational complexity while retaining a high network performance and flexibility. This paper presents ChewBaccaNN, a 0.7 mm 2 sized binary convolutional neural network (CNN) accelerator designed in GlobalFoundries 22 nm technology. By exploiting efficient data re-use, data buffering, latch-based memories, and voltage scaling, a throughput of 241 GOPS is achieved while consuming just 1.1 mW at 0.4V… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 13 publications
(11 citation statements)
references
References 23 publications
0
9
0
Order By: Relevance
“…It is clear that parallelism and data reuse (either in the form of locally buffering or by broadcasting) are the keys to amortizing the memory access cost, which is so much larger than the low-precision arithmetic cost. Techniques to mitigate these costs are to replace SRAM with low-voltage SCM, hard-wire network parameters to enable broadcasting, and use the sliding window principle (like the FMM banks in combination with the crossbar in ChewBaccaNN [1]). In essence, all these solutions boil down to designing the architecture around the data movements in a less-flexible manner.…”
Section: Comparison and Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…It is clear that parallelism and data reuse (either in the form of locally buffering or by broadcasting) are the keys to amortizing the memory access cost, which is so much larger than the low-precision arithmetic cost. Techniques to mitigate these costs are to replace SRAM with low-voltage SCM, hard-wire network parameters to enable broadcasting, and use the sliding window principle (like the FMM banks in combination with the crossbar in ChewBaccaNN [1]). In essence, all these solutions boil down to designing the architecture around the data movements in a less-flexible manner.…”
Section: Comparison and Discussionmentioning
confidence: 99%
“…A full system-on-a-chip (SoC), implemented in 22nm technology is presented including the accelerator, RISC host processor, and peripherals. • ChewBaccaNN [1] is an architecture for binary neural network inference that exploits efficient data reuse by co-designing the memory hierarchy with the neural network ran on the architecture. The hard-wired kernel size allows efficient data reuse.…”
Section: Five Low-and Mixed-precision Accelerators Reviewedmentioning
confidence: 99%
See 1 more Smart Citation
“…Replacing a real-valued MAC by XNOR-PopCount is very good for energy efficiency. E.g., in [2] a BNN accelerator is presented that achieves down to merely 4.48 fJ/Op in GF 22nm at 0.4V, where Op is a binary operation (xnor or popcount). The most common network architecture used in BNN papers is illustrated in Fig.…”
Section: Inferencementioning
confidence: 99%
“…Similar to ABCNet [26], a weighted sum is used to get to the original feature size. Compared to the per-layer ensemble this ensemble expands the network by a factor 𝑁 rather than 𝑁 2 .…”
Section: Ensemblementioning
confidence: 99%