2020
DOI: 10.48550/arxiv.2012.10873
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Sequence-to-Sequence Contrastive Learning for Text Recognition

Abstract: We propose a framework for sequence-to-sequence contrastive learning (SeqCLR) of visual representations, which we apply to text recognition. To account for the sequenceto-sequence structure, each feature map is divided into different instances over which the contrastive loss is computed. This operation enables us to contrast in a sub-word level, where from each image we extract several positive pairs and multiple negative examples. To yield effective visual representations for text recognition, we further sugg… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 36 publications
0
1
0
Order By: Relevance
“…Typical augmentation methods used include rotation, perspective and affine transformations, Gaussian noise, motion blur, resizing and padding, random or learned distortions, sharpening and cropping [39,23,24,28,1]. Proponents of STR methods select a subset of these augmentation methods to improve their models.…”
Section: Introductionmentioning
confidence: 99%
“…Typical augmentation methods used include rotation, perspective and affine transformations, Gaussian noise, motion blur, resizing and padding, random or learned distortions, sharpening and cropping [39,23,24,28,1]. Proponents of STR methods select a subset of these augmentation methods to improve their models.…”
Section: Introductionmentioning
confidence: 99%