2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR) 2017
DOI: 10.1109/icdar.2017.233
|View full text |Cite
|
Sign up to set email alerts
|

ICDAR2017 Competition on Reading Chinese Text in the Wild (RCTW-17)

Abstract: Chinese is the most widely used language in the world. Algorithms that read Chinese text in natural images facilitate applications of various kinds. Despite the large potential value, datasets and competitions in the past primarily focus on English, which bares very different characteristics than Chinese. This report introduces RCTW, a new competition that focuses on Chinese text reading. The competition features a largescale dataset with 12,263 annotated images. Two tasks, namely text localization and end-to-… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
109
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
4
2
2

Relationship

2
6

Authors

Journals

citations
Cited by 177 publications
(109 citation statements)
references
References 9 publications
0
109
0
Order By: Relevance
“…The standard evaluation protocol of MSRA-TD500 based on F-measure is used. [40] is also a long text detection dataset, consisting of 8034 training images and 4229 test images annotated with text lines. This dataset is very challenging due to very large text scale variances.…”
Section: Datasets and Evaluation Protocolsmentioning
confidence: 99%
See 2 more Smart Citations
“…The standard evaluation protocol of MSRA-TD500 based on F-measure is used. [40] is also a long text detection dataset, consisting of 8034 training images and 4229 test images annotated with text lines. This dataset is very challenging due to very large text scale variances.…”
Section: Datasets and Evaluation Protocolsmentioning
confidence: 99%
“…For oriented scene text detection on MSRA-TD500 [39] and RCTW-17 [40], we apply the same data augmentation as [20]. Besides, we also randomly rotate the images with π/2 to better handle vertical texts.…”
Section: Long Text Detection In Natural Scenesmentioning
confidence: 99%
See 1 more Smart Citation
“…:,*"()[]/' ) at the beginning and at the end of both the ground truth and the submissions are removed. For Task 2.2, the Normalized Edit Distance metric (1-N.E.D specifically, which is also used in the ICDAR 2017 competition, RCTW-17 [12]) are treated as the ranking metric. The reason of utilizing 1-N.E.D as the official ranking metric for Task 2.2 is motivated by the fact that Chinese scripts usually contain more characters than the Latin scripts, which makes word accuracy metric too harsh to evaluate Task 2.2 fairly.…”
Section: B Evaluation Metricsmentioning
confidence: 99%
“…Previous end-to-end trainable text reading models [40,14,22,28,29] only utilize images in full annotations provided by the previous benchmarks [36,3,43]. The improvement in performance of these models requires more fully annotated training data, which is extremely expensive and inefficient in annotations.…”
Section: Partially Supervised Learningmentioning
confidence: 99%