2018 16th International Conference on Frontiers in Handwriting Recognition (ICFHR) 2018
DOI: 10.1109/icfhr-2018.2018.00024
|View full text |Cite
|
Sign up to set email alerts
|

Fully Convolutional Networks for Handwriting Recognition

Abstract: Handwritten text recognition is challenging because of the virtually infinite ways a human can write the same message. Our fully convolutional handwriting model takes in a handwriting sample of unknown length and outputs an arbitrary stream of symbols. Our dual stream architecture uses both local and global context and mitigates the need for heavy preprocessing steps such as symbol alignment correction as well as complex post processing steps such as connectionist temporal classification, dictionary matching o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 21 publications
(10 citation statements)
references
References 28 publications
(46 reference statements)
0
8
0
Order By: Relevance
“…In most cases, the number of training data is several hundred or at most a few thousand for machine-learning applications in petroleum engineering [17][18][19][20][21][22][23][24][25][26][27][28], which was also noted in a previous paper [16]. Contrastingly, the number of training data is over ten thousand or even up to one million in computer-science-engineering-centered applications [29]. In spite of that, the number of training data does not necessarily guarantee reliability of training performance when inappropriate data is combined with the entire data pool.…”
Section: Introductionmentioning
confidence: 78%
“…In most cases, the number of training data is several hundred or at most a few thousand for machine-learning applications in petroleum engineering [17][18][19][20][21][22][23][24][25][26][27][28], which was also noted in a previous paper [16]. Contrastingly, the number of training data is over ten thousand or even up to one million in computer-science-engineering-centered applications [29]. In spite of that, the number of training data does not necessarily guarantee reliability of training performance when inappropriate data is combined with the entire data pool.…”
Section: Introductionmentioning
confidence: 78%
“…This task was the first tackled using LeNet [8], and is what is currently done for ideogrammatic languages such as Chinese [9] and Japanese [10]. For alphabetic languages, HTR can be also performed at word level [11], [12], [13], i.e., decoding single words that are detected in the image. This task is performed both on digitalized documents, and in scene images [14].…”
Section: Related Workmentioning
confidence: 99%
“…A fully convolutional handwriting model suggested by Petroski [28] utilizes an unknown length handwriting sample and generates an arbitrary symbol stream. Both local and global contexts are used by the dual-stream architecture and need strong pre-processing steps such as symbol alignment correction as well as complex post-processing steps such as link-time classification, dictionary matching, or language models.…”
Section: Related Workmentioning
confidence: 99%
“…The Weldegebriel [27] for classification proposes a hybrid model of two super classifiers: the CNN and the Extreme Gradient Boosting (XGBoost). CNN serves as an automatic training feature extractor for raw images for this integrated model, and XGBoost uses the extracted features as an input for recognition and classification.…”
Section: Related Workmentioning
confidence: 99%