2018
DOI: 10.1016/j.neucom.2017.09.046
|View full text |Cite
|
Sign up to set email alerts
|

FP-BNN: Binarized neural network on FPGA

Abstract: Deep neural networks (DNNs) have attracted significant attention for their excellent accuracy especially in areas such as computer vision and artificial intelligence. To enhance their performance, technologies for their hardware acceleration are being studied. FPGA technology is a promising choice for hardware acceleration, given its low power consumption and high flexibility which makes it suitable particularly for embedded systems. However, complex DNN models may need more computing and memory resources than… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
148
0
1

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 233 publications
(149 citation statements)
references
References 14 publications
0
148
0
1
Order By: Relevance
“…By comparing Wang et al [144] and Zhao et al's [158] CIFAR-10-targe ing CNN implementations with the Going Deeper [111], fpgaConvNet [142] and FP-BNN [83] ImageNet CNNs, all of which used FPGAs of similar scales, we can observe that, as precision is reduced, linear or even superlinear throughput increases can be achieved. Superlinear increases can be explained using the roo ine modelling in Section 3.…”
Section: Throughputmentioning
confidence: 99%
“…By comparing Wang et al [144] and Zhao et al's [158] CIFAR-10-targe ing CNN implementations with the Going Deeper [111], fpgaConvNet [142] and FP-BNN [83] ImageNet CNNs, all of which used FPGAs of similar scales, we can observe that, as precision is reduced, linear or even superlinear throughput increases can be achieved. Superlinear increases can be explained using the roo ine modelling in Section 3.…”
Section: Throughputmentioning
confidence: 99%
“…In [19], the authors proposed architectural changes to the Intel ALM carry-chains such that large compressors like (6:2) and (7:2) can be efficiently mapped to single ALMs. Although their proposed compressor is very efficient, for modern applications such as BNN popcounting [13], these compressors would be significantly underutilized. Similarly, Kim et.…”
Section: Related Workmentioning
confidence: 99%
“…One example of interest is that compressor trees and GPCs can be used to accelerate the XnorPopcount operations within binarized neural networks (BNNs) [1], which forms the critical path of the model's execution. BNNs enable neural networks to be utilized in resource constrained applications and can be deployed efficiently on FPGAs [13,34]; our optimizations would improve their performance further.…”
Section: Introductionmentioning
confidence: 99%
“…Since M has to be an odd number, by choosing M equal to the kernel size, which is also an odd number, and applying majority logic on the pairs placed in the same channel in a row or a column, folding is possible. Also, three is the most common kernel size for Conv layers in modern BNNs [6].…”
Section: B Xnormaj Techniquementioning
confidence: 99%