2022
DOI: 10.21203/rs.3.rs-2274499/v1
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

CNN-BiLSTM model for English Handwriting Recognition: Comprehensive Evaluation on the IAM Dataset

Abstract: We present a CNN-BiLSTM system for the problem of offline English handwriting recognition, with extensive evaluations on the public IAM dataset, including the effects of model size, data augmentation and the lexicon. Our best model achieves 3.59% CER and 9.44% WER using CNN-BiLSTM network with CTC layer.Test time augmentation with rotation and shear transformations applied to the input image, is proposed to increase recognition of difficult cases and found to reduce the word error rate by 2.5% points. We also … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 31 publications
0
1
0
Order By: Relevance
“…We assessed the performance of the text animation generation model using character error rate (CER) and word error rate (WER), which corresponds to normalized Levenshtein distances between the predicted and ground truth character sequences. Models [28][29][30] were trained and evaluated on the corresponding Aachen splits of the IAM dataset. Our approach is compared to state-of-art methods with varying characteristics, as displayed in Table 2.…”
Section: Animation Generation Experimentsmentioning
confidence: 99%
“…We assessed the performance of the text animation generation model using character error rate (CER) and word error rate (WER), which corresponds to normalized Levenshtein distances between the predicted and ground truth character sequences. Models [28][29][30] were trained and evaluated on the corresponding Aachen splits of the IAM dataset. Our approach is compared to state-of-art methods with varying characteristics, as displayed in Table 2.…”
Section: Animation Generation Experimentsmentioning
confidence: 99%
“…A primary obstacle in AHR research is the limited availability of large-scale, diverse datasets. Training deep learning models, especially those involving Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs) [10]- [13], requires ample data to learn robust representations and generalize effectively. The scarcity of labeled Arabic handwritten datasets hinders model performance and limits practical applicability.…”
Section: Introductionmentioning
confidence: 99%