2020
DOI: 10.48550/arxiv.2005.03775
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Optimizing Temporal Convolutional Network inference on FPGA-based accelerators

Abstract: Convolutional Neural Networks are extensively used in a wide range of applications, commonly including computer vision tasks like image and video classification, recognition and segmentation. Recent research results demonstrate that multilayer (deep) network involving mono-dimensional convolutions and dilation can be effectively used in time series and sequences classification and segmentation, as well as in tasks involving sequence modelling. These structures, commonly referred to as Temporal Convolutional Ne… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 28 publications
0
2
0
Order By: Relevance
“…The increasing interest in TCNs encourages the exploration of the embedded computation flow of such type of networks, such as system-level mapping to FPGAs [27], [28], or generally with dilated convolution such as in [29] and [30]. The work in [27] addresses the possibility of mapping TCNs to existing CNN accelerators, by proposing a convolution scheduling based on batch processing. The TCN architecture is based on NEURAghe [31], it consists of Sum-of-Product units, each containing a group of DSPs.…”
Section: B Related Workmentioning
confidence: 99%
“…The increasing interest in TCNs encourages the exploration of the embedded computation flow of such type of networks, such as system-level mapping to FPGAs [27], [28], or generally with dilated convolution such as in [29] and [30]. The work in [27] addresses the possibility of mapping TCNs to existing CNN accelerators, by proposing a convolution scheduling based on batch processing. The TCN architecture is based on NEURAghe [31], it consists of Sum-of-Product units, each containing a group of DSPs.…”
Section: B Related Workmentioning
confidence: 99%
“…For sparse and hybrid DNNs, Huang et al [89] propose a configurable inference engine capable of processing different sizes of sparse DNNs, while HybridDNN [90] has a hybrid architecture composed of spatial/Winograd convolution processing elements. Additionally, Carreras et al [91] present an enriched architectural template supporting efficient TCNs, together with an algorithm for optimal execution/scheduling of data-transfers to boost the implementation performance.…”
Section: Fpga-basedmentioning
confidence: 99%