2021
DOI: 10.1007/s00521-021-05813-1
|View full text |Cite
|
Sign up to set email alerts
|

Character-based handwritten text transcription with attention networks

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 20 publications
(7 citation statements)
references
References 47 publications
0
7
0
Order By: Relevance
“…An LSTM‐based GAN is proposed in Reference 26 to generate music data that sounds good. Hybrid (attentional) RNN and convolutional NN architectures have also been used for the problem of transcribing sequences of handwritten text in images 67 . In References 37,41,68–70 hybrid models are utilized to predict the trajectories of the vehicle or the vehicles in surrounding environments.…”
Section: Background and Related Workmentioning
confidence: 99%
“…An LSTM‐based GAN is proposed in Reference 26 to generate music data that sounds good. Hybrid (attentional) RNN and convolutional NN architectures have also been used for the problem of transcribing sequences of handwritten text in images 67 . In References 37,41,68–70 hybrid models are utilized to predict the trajectories of the vehicle or the vehicles in surrounding environments.…”
Section: Background and Related Workmentioning
confidence: 99%
“…Contrary to the CRNN-CTC architecture, attention-based models learn to align image pixels with the target sequence. As a result, the network learns to focus on a small relevant part of the feature vector to predict each token [19,22]. Another strength of this architecture is that the recurrent decoder learns an implicit language model at character-level.…”
Section: Handwriting Recognitionmentioning
confidence: 99%
“…It recently gained popularity for speech recognition, image captioning and neural translation [3,8,29]. The architecture has since been adapted for HTR [19,22], as illustrated in Fig. 2.…”
Section: The Attention-based Seq2seq Architecturementioning
confidence: 99%
“…The NN architecture based on convolutional and 1D-LSTM layers is able to learn similar features with a significantly smaller computational cost [45]. Some notable state-of-the-art systems are only made up of CNN layers or attention techniques without any recurrent layer [13,42,43,44,52,57,58].…”
Section: Related Workmentioning
confidence: 99%