2019
DOI: 10.1109/tvlsi.2019.2941250
|View full text |Cite
|
Sign up to set email alerts
|

High-Performance CNN Accelerator on FPGA Using Unified Winograd-GEMM Architecture

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
38
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 86 publications
(42 citation statements)
references
References 22 publications
0
38
0
Order By: Relevance
“…The authors of [9][10][11][12] proposed to accelerate CNN on FPGA using simplified numerical precision to save chip resource consumption. The authors of [13,14] proposed CNN architecture implemented in FPGA with the Winograd algorithm to reduce the complexity of convolution operation and accelerate the computation process. Bai et al [15] specifically used depthwise separable convolution to implement the CNN accelerator.…”
Section: Related Workmentioning
confidence: 99%
“…The authors of [9][10][11][12] proposed to accelerate CNN on FPGA using simplified numerical precision to save chip resource consumption. The authors of [13,14] proposed CNN architecture implemented in FPGA with the Winograd algorithm to reduce the complexity of convolution operation and accelerate the computation process. Bai et al [15] specifically used depthwise separable convolution to implement the CNN accelerator.…”
Section: Related Workmentioning
confidence: 99%
“…CNNs extract important features embedded in the input data and are increasingly computationally efficient. As recent studies have shown the effectiveness of FPGA as a hardware accelerator for the CNNs [51][52][53], the CNN in this system is to be built on FPGA as a real-time and low power consumption system. The CNN is built using Theano [54], and it consists of two convolutional layers, two pooling layers, one all-to-all connection layer and one output layer, as shown in Figure 1F.…”
Section: Regression Neural Networkmentioning
confidence: 99%
“…Winograd filtering is a known technique to reduce the number of multiplications of a convolution. The technique was efficiently implemented on FPGA [147][148][149][150].…”
Section: Hardware-oriented Deep Neural Network Optimizationsmentioning
confidence: 99%
“…The main data quantization and data reduction techniques are summarized in Table 4. [146][147][148][149][150] Data reduction techniques are normally applied together with data quantization. Together, they generate very efficient solutions with a small accuracy reduction when compared to solutions without optimizations.…”
Section: Hardware-oriented Deep Neural Network Optimizationsmentioning
confidence: 99%