Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2021
DOI: 10.18653/v1/2021.naacl-main.427
|View full text |Cite
|
Sign up to set email alerts
|

Supporting Clustering with Contrastive Learning

Abstract: Unsupervised clustering aims at discovering the semantic categories of data according to some distance measured in the representation space. However, different categories often overlap with each other in the representation space at the beginning of the learning process, which poses a significant challenge for distance-based clustering in achieving good separation between different categories. To this end, we propose Supporting Clustering with Contrastive Learning (SCCL) -a novel framework to leverage contrasti… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
38
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 86 publications
(53 citation statements)
references
References 35 publications
0
38
0
Order By: Relevance
“…However, generators are computationally expensive for end-to-end training, and often less effective than the discriminative models [15,27,33] in feature learning [11]. Recent research has considered contrastive learning in clustering [28,52,70,85]. We discuss the drawback of them and their relations to TCC in Sec.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…However, generators are computationally expensive for end-to-end training, and often less effective than the discriminative models [15,27,33] in feature learning [11]. Recent research has considered contrastive learning in clustering [28,52,70,85]. We discuss the drawback of them and their relations to TCC in Sec.…”
Section: Related Workmentioning
confidence: 99%
“…This has further motivated the development of a two-stage clustering pipeline [70] with contrastive pre-training and k-means [55]. An alternative simple migration [85] yields a composition of an InfoNCE loss [61] and a clustering one [77]. Compared with the deep generative counterparts [16,36,53,82], contrastive clustering is free from decoding and computationally practical, with guaranteed feature quality.…”
Section: Introductionmentioning
confidence: 99%
“…However, this requires data augmentation to create positive example pairs. For text, some augmentations use back-translation (Cao and Wang, 2021;Zhang et al, 2021b). Taking inspiration from these clustering and representation learning techniques, we employ back-translation as data augmentation to create more positive pairs, improving the learning of attention weights between event mentions.…”
Section: Back-translationmentioning
confidence: 99%
“…Several new unsupervised deep clustering approaches use contrastive loss for clustering images Zhong et al, 2020) and text (Zhang et al, 2021b). These methods require data augmentation to create positive example pairs.…”
Section: Related Workmentioning
confidence: 99%
“…We name our approach Pairwise Supervised Contrastive Learning (PairSupCon). As noticed by the recent work (Wu et al, 2018;Zhang et al, 2021), instance discrimination learning can implicitly group similar instances together in the representation space without any explicit learning force directs to do so. PairSupCon leverages this implicit grouping effect to bring together representations from the same semantic category while, simultaneously enhancing the semantic entailment and contradiction reasoning capability of the model.…”
Section: Introductionmentioning
confidence: 99%