2022
DOI: 10.48550/arxiv.2201.12498
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Investigating Why Contrastive Learning Benefits Robustness Against Label Noise

Abstract: Self-supervised contrastive learning has recently been shown to be very effective in preventing deep networks from overfitting noisy labels. Despite its empirical success, the theoretical understanding of the effect of contrastive learning on boosting robustness is very limited. In this work, we rigorously prove that the representation matrix learned by contrastive learning boosts robustness, by having: (i) one prominent singular value corresponding to every sub-class in the data, and remaining significantly s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 24 publications
(50 reference statements)
0
1
0
Order By: Relevance
“…These pretrained, generalizable encoders have become a popular molecular design tool in recent years. However, these models may operate on different chemical representations with no clear optimal choice. Contrastive learning approaches are able to integrate several data modalities, can boost robustness on downstream tasks, and have been shown to be successful in multiple fields. We explore a scheme that uses contrastive learning of multiple molecular modalities, and our experiments show that this strategy leads to broadly applicable and robust representations. More generally, we seek a generative foundation model of small molecules that decouples conditional generation from fine-tuning of the foundation model and provides a path forward for future multimodal representation learning advances.…”
Section: Introductionmentioning
confidence: 98%
“…These pretrained, generalizable encoders have become a popular molecular design tool in recent years. However, these models may operate on different chemical representations with no clear optimal choice. Contrastive learning approaches are able to integrate several data modalities, can boost robustness on downstream tasks, and have been shown to be successful in multiple fields. We explore a scheme that uses contrastive learning of multiple molecular modalities, and our experiments show that this strategy leads to broadly applicable and robust representations. More generally, we seek a generative foundation model of small molecules that decouples conditional generation from fine-tuning of the foundation model and provides a path forward for future multimodal representation learning advances.…”
Section: Introductionmentioning
confidence: 98%