Design, Automation &Amp; Test in Europe Conference &Amp; Exhibition (DATE), 2017 2017
DOI: 10.23919/date.2017.7927210
|View full text |Cite
|
Sign up to set email alerts
|

Hardware architecture of Bidirectional Long Short-Term Memory Neural Network for Optical Character Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
50
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
3
3
2

Relationship

2
6

Authors

Journals

citations
Cited by 41 publications
(50 citation statements)
references
References 8 publications
0
50
0
Order By: Relevance
“…RNN needs to run for multiple time-steps for each task to be completed. The computation of the recurrent unit can be unrolled over timesteps [92]. However, this cannot be fully parallelized, as discussed earlier.…”
Section: ) Compute-specificmentioning
confidence: 99%
“…RNN needs to run for multiple time-steps for each task to be completed. The computation of the recurrent unit can be unrolled over timesteps [92]. However, this cannot be fully parallelized, as discussed earlier.…”
Section: ) Compute-specificmentioning
confidence: 99%
“…Ferreira et al proposed an FPGA accelerator of LSTM in [7] for a learning problem of adding two 8-bit numbers with weights stored in on-chip memory. Rybalkin et al [8] presented the first hardware architecture designed for BiLSTM for OCR. The architecture was implemented with 5-bit fixed-point numbers for weights and activations which were stored in on-chip memory.…”
Section: B Related Workmentioning
confidence: 99%
“…There has been previous work [7,8,9,10] with FPGA based implementations such that all the weights are stored in the on-chip memory, but this is expensive and limits the size of models that can be deployed. When the RNN model is too large that the weights need to be stored on an external DRAM, it is not efficient because the fetched weights are typically used only once for each output computation.…”
Section: Introductionmentioning
confidence: 99%
“…An option are FPGAs, which allow to design a specialized hardware architecture for DNNs, but at much less effort than building a computer chip from scratch. There are several examples of FPGA implementations dealing with redundancy in DNNs [5,6,7,8,9,10]. FPGAs consume little energy, therefore they are good candidates for embedded applications.…”
Section: Related Workmentioning
confidence: 99%