2019 32nd International Conference on VLSI Design and 2019 18th International Conference on Embedded Systems (VLSID) 2019
DOI: 10.1109/vlsid.2019.00055
|View full text |Cite
|
Sign up to set email alerts
|

UniWiG: Unified Winograd-GEMM Architecture for Accelerating CNN on FPGAs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 20 publications
(8 citation statements)
references
References 11 publications
0
8
0
Order By: Relevance
“…FPGA implementations of Winograd convolution are presented in [81], which incorporate feature map caching using linebuff structure, data reuse, effective use of pipelining for PEs, and parallel processing of convolution operations. A unified architecture incorporating the Winograd and general matrix multiplication (GEMM) named UniWig is presented in [82]. Instead of using separate PEs for convolution and dense layers, UniWig utilizes the same set of PEs and blocked Winograd filtering to ensure proper resource utilization.…”
Section: ) Optimized Convolution In Cnnmentioning
confidence: 99%
“…FPGA implementations of Winograd convolution are presented in [81], which incorporate feature map caching using linebuff structure, data reuse, effective use of pipelining for PEs, and parallel processing of convolution operations. A unified architecture incorporating the Winograd and general matrix multiplication (GEMM) named UniWig is presented in [82]. Instead of using separate PEs for convolution and dense layers, UniWig utilizes the same set of PEs and blocked Winograd filtering to ensure proper resource utilization.…”
Section: ) Optimized Convolution In Cnnmentioning
confidence: 99%
“…As shown in Figure 6, according to different design concepts and requirements, FPGA-based neural network optimization technology can be roughly divided into optimization for data and operation, optimization for bandwidth, and optimization for memory and access, among others, which are introduced in detail below. [71][72][73][74][75][76][77][78], less computations [79][80][81], improve calculation speed [82][83][84][85], Winograd fast convolution algorithm [86][87][88][89][90][91], Im2col convolution optimization algorithm [92][93][94][95][96][97], pipelined design [98][99][100][101][102], Roofline model [103][104][105], ping-pong cache [106][107][108][109], input feature map reuse [110,111], filter reuse [111,112], convolutional reuse [110]…”
Section: Neural Network Optimization Technology Based On Fpgamentioning
confidence: 99%
“…In this paper, we advocated a paradigm in which an equal number of processing elements might be used to boost both Winograd-based convolution and GEMM [7]. The PE contains a multipliers adder, register, and FIFO'S.…”
Section: Hardware Implementationmentioning
confidence: 99%
“…Convolutions are reduced to generic element-wise matrix multiplications (GEMMs) in the usual method [7]. Convolution based on the Fast Fourier transform (FFT) is less computationally complex than the traditional technique.…”
Section: Introduction and Literature Surveymentioning
confidence: 99%