Proceedings of the 2017 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays 2017
DOI: 10.1145/3020078.3021741
|View full text |Cite
|
Sign up to set email alerts
|

Accelerating Binarized Convolutional Neural Networks with Software-Programmable FPGAs

Abstract: Convolutional neural networks (CNN) are the current stateof-the-art for many computer vision tasks. CNNs outperform older methods in accuracy, but require vast amounts of computation and memory. As a result, existing CNN applications are typically run on clusters of CPUs or GPUs. Research on FPGA acceleration of CNN workloads has achieved reductions in power and energy consumption. However, large GPUs outperform modern FPGAs in throughput, and the existence of compatible deep learning frameworks give GPUs a si… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
143
0
5

Year Published

2017
2017
2022
2022

Publication Types

Select...
4
4
2

Relationship

0
10

Authors

Journals

citations
Cited by 329 publications
(151 citation statements)
references
References 17 publications
0
143
0
5
Order By: Relevance
“…sign in Github to realize on the Xilinx PYNQ board which has the same FPGA used in other designs. From Table 7, compared with Zhao's implementation [34], the classification accuracy was almost the same, as for the performance per power efficiency (FPS/Watt), it is 5.18 times better, Compared with the FINN, the memory efficiency was 3.98 time better, and the performance per power efficiency is almost the same. Thus, our design achieves power efficiency CNN, since all the circuits are operated on chip primitives.…”
Section: Compared With An Edge Pruning Methodsmentioning
confidence: 90%
“…sign in Github to realize on the Xilinx PYNQ board which has the same FPGA used in other designs. From Table 7, compared with Zhao's implementation [34], the classification accuracy was almost the same, as for the performance per power efficiency (FPS/Watt), it is 5.18 times better, Compared with the FINN, the memory efficiency was 3.98 time better, and the performance per power efficiency is almost the same. Thus, our design achieves power efficiency CNN, since all the circuits are operated on chip primitives.…”
Section: Compared With An Edge Pruning Methodsmentioning
confidence: 90%
“…Zhao et al [22] propose BNN, an FPGA implementation of NN-128 with binary weights on board ZedBoard. They focus on accelerating the neural network in a very reduced FPGA, so the resulting throughput is very low.…”
Section: R E L At E D W O R Kmentioning
confidence: 99%
“…In Section 7, we validate the performance of our filter matrix packing algorithm with an FPGA implementation. Additionally, we compare our implementation to previous state-of-the-art FPGA results [57,43,16,70]. Figure 2 compares standard CNNs to two recent CNN variants, separable convolution [12,25] and shift convolution [65], as shown in Figure 2.…”
Section: Asic and Fpga Accelerators For Cnnsmentioning
confidence: 99%