2017
DOI: 10.48550/arxiv.1703.08289
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Deep Direct Regression for Multi-Oriented Scene Text Detection

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
40
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 38 publications
(40 citation statements)
references
References 15 publications
0
40
0
Order By: Relevance
“…2, such as [2,11,23,24,29,32]. As its name described, the direct regression method directly calculates the error between the prediction and ground truth [7,8,35].…”
Section: Rbox Regression Parametersmentioning
confidence: 99%
“…2, such as [2,11,23,24,29,32]. As its name described, the direct regression method directly calculates the error between the prediction and ground truth [7,8,35].…”
Section: Rbox Regression Parametersmentioning
confidence: 99%
“…The aspect ratio of text lines varies greatly, and limited anchors cannot cover the size or aspect ratio of all objects; thus, many methods are anchor-free. Both [4] and [1] generate labels with shrunk segmentation maps, and regress the vertices or angles of the bounding box on positive pixels. [29] generates a corner map and a position-sensitive segmentation map, calculates oriented bounding boxes based on the corner map, and calculates the score for each bounding box using the position-sensitive segmentation map.…”
Section: B Oriented Objects Detectionmentioning
confidence: 99%
“…[23] classifies text line and regresses its location with different feature which achieves significant improvement on oriented text line. [15] and [48] investigate to generate shrinked text line segmentation map then regress text sides or vertexes on text center.…”
Section: Related Workmentioning
confidence: 99%
“…Following [6,29,28] we resize images to 1280 × 768 in inference and report the single-scale result. We compare our method with other state-of-the-art meth-Methods P R F FPS Zhang et al [45] 71 43 54 -SegLink [35] 73.1 76.8 75.0 -EAST [48] 83.57 73.47 78.20 13.2 EAST [48] 83.27 78.33 80.72 -He et al [15] 82 80 81 -PixelLink [6] 85.5 82.0 83.7 3.0 Lyu et al [29] 89.5 79.7 84.3 1 TextSnake [28] 84.9 ods and show results in Table 3. Our method achieves better performance (precision: 88.51%, recall: 84.16% and Fmeasure: 86.28%) compared with other segmentation based methods [6,29].…”
Section: Experiments On Icdar2015mentioning
confidence: 99%