2019
DOI: 10.1587/transinf.2018rcp0008
|View full text |Cite
|
Sign up to set email alerts
|

RNA: An Accurate Residual Network Accelerator for Quantized and Reconstructed Deep Neural Networks

Abstract: With the continuous refinement of Deep Neural Networks (DNNs), a series of deep and complex networks such as Residual Networks (ResNets) show impressive prediction accuracy in image classification tasks. Unfortunately, the structural complexity and computational cost of residual networks make hardware implementation difficult. In this paper, we present the quantized and reconstructed deep neural network (QR-DNN) technique, which first inserts batch normalization (BN) layers in the network during training, and … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
3
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 11 publications
(3 citation statements)
references
References 19 publications
0
3
0
Order By: Relevance
“…The rapid development of DNNs in recent years has led them to become a core enabler for a broad spectrum of application areas, such as computer vision, natural language processing, medical engineering, autonomous driving, and virtual reality. Meanwhile, there is a rising trend showing that the deployment of these applications is shifting from traditional cloud computing platforms (e.g., servers and supercomputers) to edge devices (e.g., mobile and handheld platforms) whose power efficiency is one of the major constraints [148][149][150][151][152][153][154]. Field Programmable Gate Arrays (FPGAs) and mobile devices, as the two most popular substrates in edge devices 1 , have established their dominance through the delivery of promising energy/power efficiency and performance.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…The rapid development of DNNs in recent years has led them to become a core enabler for a broad spectrum of application areas, such as computer vision, natural language processing, medical engineering, autonomous driving, and virtual reality. Meanwhile, there is a rising trend showing that the deployment of these applications is shifting from traditional cloud computing platforms (e.g., servers and supercomputers) to edge devices (e.g., mobile and handheld platforms) whose power efficiency is one of the major constraints [148][149][150][151][152][153][154]. Field Programmable Gate Arrays (FPGAs) and mobile devices, as the two most popular substrates in edge devices 1 , have established their dominance through the delivery of promising energy/power efficiency and performance.…”
Section: Introductionmentioning
confidence: 99%
“…On FPGAs, weight quantization is a natural fit. Besides storage reduction, the additional benefits include (1) the DSP on FPGA can support multiple multiply-and-accumulate (MAC) computations with appropriate weight (and activation) quantization, and (2) the look-up table (LUT) computing resources can support low-precision computing [150,[178][179][180][181][182][183][184][185]. Low-bit-width fixedpoint quantization is achieved in [179] through greedy solution to determine the radix position of each layer for quantization, and in [182] with a hybrid quantization scheme that allows different bit-widths for weights to provide more flexibility.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation