2021
DOI: 10.1109/tvlsi.2021.3109580
|View full text |Cite
|
Sign up to set email alerts
|

A Reconfigurable Neural Network Processor With Tile-Grained Multicore Pipeline for Object Detection on FPGA

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(5 citation statements)
references
References 29 publications
0
5
0
Order By: Relevance
“…CNFG [2] CNFG [3] CNFG [4] CNFG [5] CNFG [6] CNFG [7] CNFG [8] CNFG [9] CNFG [10] CNFG [11] CNFG [12] CNFG [13] CNFG [14] CNFG [15] In [0] In [1] In [2] In [3] Out Programmable Filter…”
Section: Figmentioning
confidence: 99%
See 1 more Smart Citation
“…CNFG [2] CNFG [3] CNFG [4] CNFG [5] CNFG [6] CNFG [7] CNFG [8] CNFG [9] CNFG [10] CNFG [11] CNFG [12] CNFG [13] CNFG [14] CNFG [15] In [0] In [1] In [2] In [3] Out Programmable Filter…”
Section: Figmentioning
confidence: 99%
“…In particular, it has been demonstrated in [3] that is possible to achieve a complete design flow for mapping CNN on FPGA. The FPGA reconfigurability provides the advantage of adapting their design to the CNN inference models without requiring significant modification of the hardware architecture [14]. Besides, FPGAs achieve extensive computational parallelism, enabling the usage of depth-wise separable convolution instead of standard convolution CNN, reducing the number of used Multiply and Accumulate (MAC) modules [15].…”
Section: Related Workmentioning
confidence: 99%
“…Gong et al [35] proposed an accelerator architecture that used both static and dynamic reconfigurabilities of the hardware. Chang et al [36] proposed a reconfigurable CNN processor with a parallelism complementary and hierarchical multicore pipelining architecture. By well-designed architecture and data path, these works' throughput achieved over 1000 GOPS, but they did not pay attention to reducing latency.…”
Section: Cnn Fpga Implementationmentioning
confidence: 99%
“…Other works [14]- [16] used on-chip memory for low off-chip memory access. Pipelined architecture [27], [28] has been explored to speed up infer-ence time. Although these works can improve the performance of the accelerator, they still use a double data rate (DDR) for the off-chip memory.…”
Section: Related Workmentioning
confidence: 99%