2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR) 2017
DOI: 10.1109/icdar.2017.237
|View full text |Cite
|
Sign up to set email alerts
|

ICDAR2017 Robust Reading Challenge on Multi-Lingual Scene Text Detection and Script Identification - RRC-MLT

Abstract: HAL is a multi-disciplinary open access archive for the deposit and dissemination of scientific research documents, whether they are published or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers. L'archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d'enseignement et de

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
216
0

Year Published

2018
2018
2019
2019

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 350 publications
(222 citation statements)
references
References 9 publications
0
216
0
Order By: Relevance
“…Our CharNet is evaluated on three standard benchmarks: ICDAR 2015 [17], Total-Text [5], and ICDAR MLT 2017 [27]. ICDAR 2015 includes 1,500 images collected by using Google Glasses.…”
Section: Experiments Results and Comparisonsmentioning
confidence: 99%
“…Our CharNet is evaluated on three standard benchmarks: ICDAR 2015 [17], Total-Text [5], and ICDAR MLT 2017 [27]. ICDAR 2015 includes 1,500 images collected by using Google Glasses.…”
Section: Experiments Results and Comparisonsmentioning
confidence: 99%
“…To validate its effectiveness, we adopt the state-of-the-art RetinaNet [19] as our baseline model and present a simple and intuitive text detector named STELA (Scene TExt Detector with Learned Anchor), in which each location of feature maps only associates with one anchor. Following the standard evaluation protocols in each benchmark, our method achieves comparable performances with an F-measure 0.887 on ICDAR 2013 [12], 0.833 on ICDAR 2015 [11] and 0.715 on ICDAR 2017 MLT [25]. Besides, our method is a totally real-time scene text detector with 26.5f ps at 800p, which surpasses all of anchor-based methods.…”
Section: Introductionmentioning
confidence: 86%
“…Restricted by the hardware, the batch size is set to 4 and the initial learning rate is set to 10 −4 . We randomly pick up 100,000 images from SynthText [7] to pretrain the network for 5 epochs, and collect real data from ICDAR 2013 [12], 2015 [11] and 2017 [25] to finetune a final model for 25 epochs. The learning rate is decayed to 10 −5 after 15 epochs of finetuning.…”
Section: Implementation Detailsmentioning
confidence: 99%
“…CVSI2015 [25] is released for the ICDAR 2015 Competition on Video Script Identification, containing text line images of 10 Indian scripts. RRC-MLT2017 [26] is released for ICDAR 2017 Competition on MLT-Task2, comprising 68,613 training, 16,255 validation and 97,619 test cropped images. This dataset holds an extremely imbalanced distribution among 7 scripts and especially tilts to Latin.…”
Section: Methodsmentioning
confidence: 99%
“…2) We design softermax loss to accomplish patch-level weak supervision on local predictions with image-level label. 3) Experiments are conducted on three public datasets, i.e., SIW-13 [24], CVSI2015 [25] and RRC-MLT2017 [26], and achieve state-of-the-art performance.…”
Section: Introductionmentioning
confidence: 99%