2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2021
DOI: 10.1109/cvpr46437.2021.01505
|View full text |Cite
|
Sign up to set email alerts
|

Sequence-to-Sequence Contrastive Learning for Text Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
36
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 87 publications
(46 citation statements)
references
References 38 publications
0
36
0
Order By: Relevance
“…Inspired by success of contrastive learning in other domains, [7] proposed contrastive predictive coding which learned representations by predicting the future in the latent space and showed great advances in various speech recognition task. Also, [16] extended SimCLR model [10] to EEG data. More recently [19] proposed multitask contrastive learning approach which capture temporal and contextual information from time-series.…”
Section: Self-supervised Learning For Time-seriesmentioning
confidence: 99%
See 1 more Smart Citation
“…Inspired by success of contrastive learning in other domains, [7] proposed contrastive predictive coding which learned representations by predicting the future in the latent space and showed great advances in various speech recognition task. Also, [16] extended SimCLR model [10] to EEG data. More recently [19] proposed multitask contrastive learning approach which capture temporal and contextual information from time-series.…”
Section: Self-supervised Learning For Time-seriesmentioning
confidence: 99%
“…(2) They may not able to capture low and high frequency time varying features which is important given characteristics of time-series data [12], [13]. More recently, some work on contrastive learning for EEG, ECG and time-series has been done but they are mostly data or application specific [14], [15], [16].…”
Section: Introductionmentioning
confidence: 99%
“…Most of these methods follow the "pre-training and fine-tuning" paradigm. However, this learning paradigm has been rarely studied in the STR field and only a few methods (Aberdam et al 2021;Chen et al 2020) have been proposed.…”
Section: Introductionmentioning
confidence: 99%
“…Scripts of the recognized textual content are detected by a CNN, whereas for detection, they proposed an UrduNet, an integration of CNN and LSTM networks. Aberdam et al [13] proposed architecture for SeqCLR of visual depictions that they employ for recognizing text. To demonstrate the sequence-to-sequence framework, every feature map is separated into different cases where the contrastive loss is calculated.…”
Section: Introductionmentioning
confidence: 99%