The 2021 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays 2021
DOI: 10.1145/3431920.3439296
|View full text |Cite
|
Sign up to set email alerts
|

FracBNN: Accurate and FPGA-Efficient Binary Neural Networks with Fractional Activations

Abstract: Binary neural networks (BNNs) have 1-bit weights and activations. Such networks are well suited for FPGAs, as their dominant computations are bitwise arithmetic and the memory requirement is also significantly reduced. However, compared to start-of-the-art compact convolutional neural network (CNN) models, BNNs tend to produce a much lower accuracy on realistic datasets such as Im-ageNet. In addition, the input layer of BNNs has gradually become a major compute bottleneck, because it is conventionally excluded… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4

Citation Types

0
40
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
3
2

Relationship

2
8

Authors

Journals

citations
Cited by 74 publications
(40 citation statements)
references
References 43 publications
(71 reference statements)
0
40
0
Order By: Relevance
“…Compared with CPUs and GPUs, FPGAs have the unique capability of customizing the control flow and data paths, which has demonstrated tremendous potential in various application domains, including stencil computations [7,8,19,38], neural networks [33,52,56], and general graph algorithms [5,28,55]. This makes the FPGA a naturally good candidate platform for SSSP acceleration, since the high-throughput on-chip priority queues [2,36] enable effective control over the trade-off between parallelism and the amount of work [1,35].…”
Section: Introductionmentioning
confidence: 99%
“…Compared with CPUs and GPUs, FPGAs have the unique capability of customizing the control flow and data paths, which has demonstrated tremendous potential in various application domains, including stencil computations [7,8,19,38], neural networks [33,52,56], and general graph algorithms [5,28,55]. This makes the FPGA a naturally good candidate platform for SSSP acceleration, since the high-throughput on-chip priority queues [2,36] enable effective control over the trade-off between parallelism and the amount of work [1,35].…”
Section: Introductionmentioning
confidence: 99%
“…We implement HiSparse using high-level synthesis (HLS). While recent years have seen a rapidly increasing adoption of HLS for accelerator development, a majority of existing HLS designs target dense computations, such as dense matrix multiplication [9][10][11], image/video processing [12][13][14], and convolutional neural networks [15][16][17]. Developing high-performance sparse accelerators using HLS is more challenging because the irregular compute pattern of sparse workloads causes bank conflicts and carried dependencies.…”
Section: Introductionmentioning
confidence: 99%
“…Field Programmable Gate Arrays (FPGAs) are increasingly prominent in modern heterogeneous computer systems. Specialized hardware designs provide unprecedented efficiency in domains such as machine learning [74,83,101,102,122,127,128], compression [92,125], database operations [88,96,104], graph processing [36,47,112,129], networking [41,52,111], and storage virtualization [78]. To realize the benefits of FPGAs, systems researchers have built operating systems [53,73,77,106], virtualization support [42,46,80,85,113,120,123,124], just-in-time compilers [97], and high-level synthesis tools [43,44,61,116,117].…”
Section: Introductionmentioning
confidence: 99%