2019 International Conference on Document Analysis and Recognition (ICDAR) 2019
DOI: 10.1109/icdar.2019.00-16
|View full text |Cite
|
Sign up to set email alerts
|

End-to-End Measure for Text Recognition

Abstract: Measuring the performance of text recognition and text line detection engines is an important step to objectively compare systems and their configuration. There exist wellestablished measures for both tasks separately. However, there is no sophisticated evaluation scheme to measure the quality of a combined text line detection and text recognition system. The F-measure on word level is a well-known methodology, which is sometimes used in this context. Nevertheless, it does not take into account the alignment o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(8 citation statements)
references
References 6 publications
(15 reference statements)
0
8
0
Order By: Relevance
“…The RO provided by reference transcripts and/or other layout GT annotations is generally only one among several possible RO annotations which would be all correct. Therefore, mixing RO and word recognition errors into a single assessment measure (as in [14,25]) does mot seem the best idea for understanding which are the inner issues of an end-to-end full-page HTR system.…”
Section: The Reading Order Problemmentioning
confidence: 99%
See 2 more Smart Citations
“…The RO provided by reference transcripts and/or other layout GT annotations is generally only one among several possible RO annotations which would be all correct. Therefore, mixing RO and word recognition errors into a single assessment measure (as in [14,25]) does mot seem the best idea for understanding which are the inner issues of an end-to-end full-page HTR system.…”
Section: The Reading Order Problemmentioning
confidence: 99%
“…Finally, both measures are somehow combined to obtain a single scalar figure which hopefully represents an "overall performance" metric [14]. In a similar vein, but explicitly devoted to HTR evaluation, the work presented in [25],goes deeper in the metric combination idea, with daunting mathematical formulation. However, this is a utterly theoretical work which does not provide any empirical evidence that would support the proposed formulation or methods in practice.…”
Section: Integrating Evaluation Of Wer and Reading Order Mismatchmentioning
confidence: 99%
See 1 more Smart Citation
“…CER is the inverted accuracy, and defined as CER = (i + s + d) / n, where n is the total number of characters, i the minimal number of character insertions, s the substitutions and d the deletions required to transform the reference text into the OCR output. 6 [16] propose an "end-to-end measure" which is based on the CER, but with alignment between GT and OCR results in a way that makes it configurable whether differences in the reading order or the over-/under-segmentation of text lines are penalized.…”
Section: State-of-the-artmentioning
confidence: 99%
“…The difficulties underlying the evaluation of page-level HTR results boil down to a Reading Order (RO) problem [7,26,30,33] . A number of recent proposals try to heuristically weight and combine both word recognition and LA geometric errors into a single scalar value [10,19] . Unfortunately, this hinders the capability to sort out the nature of the corresponding errors and thereby making a comprehensive, useful assessment.…”
Section: Introductionmentioning
confidence: 99%