2020 30th International Conference on Field-Programmable Logic and Applications (FPL) 2020
DOI: 10.1109/fpl50879.2020.00055
|View full text |Cite
|
Sign up to set email alerts
|

LogicNets: Co-Designed Neural Networks and Circuits for Extreme-Throughput Applications

Abstract: Deployment of deep neural networks for applications that require very high throughput or extremely low latency is a severe computational challenge, further exacerbated by inefficiencies in mapping the computation to hardware. We present a novel method for designing neural network topologies that directly map to a highly efficient FPGA implementation. By exploiting the equivalence of artificial neurons with quantized inputs/outputs and truth tables, we can train quantized neural networks that can be directly co… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
30
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 65 publications
(30 citation statements)
references
References 15 publications
0
30
0
Order By: Relevance
“…The same group have also developed a library for quantization-aware training, Brevitas [46], based on PyTorch model formats. The Log-icNets design flow [47], also from Xilinx Research Labs, allows for the training of quantized DNNs that map to highly efficient Xilinx FPGA implementations. A comparison between the approach presented here and Logic-Nets is provided in Section VII.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…The same group have also developed a library for quantization-aware training, Brevitas [46], based on PyTorch model formats. The Log-icNets design flow [47], also from Xilinx Research Labs, allows for the training of quantized DNNs that map to highly efficient Xilinx FPGA implementations. A comparison between the approach presented here and Logic-Nets is provided in Section VII.…”
Section: Related Workmentioning
confidence: 99%
“…Further, we compare the results obtained using the QKeras and hls4ml workflow to LogicNets [47]; another work on extreme low-latency, low-resource, fully-unfolded (II=1) FPGA implementations. The metrics are those quoted in Table III.…”
Section: Ultra Low-latency Quantized Model On Fpga Hardwarementioning
confidence: 99%
“…Since, each unit requires only one BRAM access, it is faster than FINN-R. Also, it requires lower power, since our network requires fewer BRAMs than FINN-R. Note that this comparison excludes the softmax part [20].…”
Section: Comparison With Neural Networkmentioning
confidence: 99%
“…With LUTNet, we reported area efficiency improvements of around 2× over ReBNet [3], the state-of-the-art BNN at the time, for problems of widely varying scale. More recent tools, including NullaNet [12] and LogicNets [18], also generate small LUTs as core components, but LUTNet remains unique in directly exposing a netlist's LUTs as differentiable functions trainable via stochastic gradient descent.…”
Section: Introductionmentioning
confidence: 99%
“…RELATED WORK 2.1 FPGA-Tailored DNN ArchitecturesLUT-based DNN inference accelerators have been shown to achieve remarkable performance when deployed on FPGAs. NullaNet[12] and LogicNets[18] were conceived with small-scale classification tasks in mind, for which they reached latency in the tens of nanoseconds and throughput in the hundreds of millions of samples per second. Going beyond FPGA-tailored network design, our previously proposed LUTNet topologies can be trained via stochastic gradient descent[20].…”
mentioning
confidence: 99%