2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2021
DOI: 10.1109/cvpr46437.2021.01185
|View full text |Cite
|
Sign up to set email alerts
|

Scene Text Telescope: Text-Focused Scene Image Super-Resolution

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
40
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 71 publications
(53 citation statements)
references
References 33 publications
1
40
0
Order By: Relevance
“…Besides, as shown in Figure 7, the scene dataset comprises many vertical text images, which indeed put obstacles to those baselines (e.g., CRNN [59], ASTER [60], and MORAN [45]) that simply transform original input to 1-D feature sequence. By contrast, those 2-D methods (e.g., SAR [36], SRN [88] and TransOCR [10]) achieve better performance on this dataset, as 2-D feature maps are more robust to tackle text images with special layouts (e.g., vertical or curved). Further, by taking advantage of the self-attention modules, TransOCR [10] surpasses all its counterparts with recognition accuracy 63.3% as it is capable of modeling the sequential features more flexibly.…”
Section: Analysis Of Experimental Resultsmentioning
confidence: 85%
See 4 more Smart Citations
“…Besides, as shown in Figure 7, the scene dataset comprises many vertical text images, which indeed put obstacles to those baselines (e.g., CRNN [59], ASTER [60], and MORAN [45]) that simply transform original input to 1-D feature sequence. By contrast, those 2-D methods (e.g., SAR [36], SRN [88] and TransOCR [10]) achieve better performance on this dataset, as 2-D feature maps are more robust to tackle text images with special layouts (e.g., vertical or curved). Further, by taking advantage of the self-attention modules, TransOCR [10] surpasses all its counterparts with recognition accuracy 63.3% as it is capable of modeling the sequential features more flexibly.…”
Section: Analysis Of Experimental Resultsmentioning
confidence: 85%
“…By contrast, those 2-D methods (e.g., SAR [36], SRN [88] and TransOCR [10]) achieve better performance on this dataset, as 2-D feature maps are more robust to tackle text images with special layouts (e.g., vertical or curved). Further, by taking advantage of the self-attention modules, TransOCR [10] surpasses all its counterparts with recognition accuracy 63.3% as it is capable of modeling the sequential features more flexibly. At last, we notice that SEED [54] does not perform well on this dataset.…”
Section: Analysis Of Experimental Resultsmentioning
confidence: 85%
See 3 more Smart Citations