2022
DOI: 10.48550/arxiv.2211.04928
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

miCSE: Mutual Information Contrastive Learning for Low-shot Sentence Embeddings

Abstract: This paper presents miCSE, a mutual information-basedContrastive learning framework that significantly advances the state-of-the-art in few-shot sentence embedding. The proposed approach imposes alignment between the attention pattern of different views during contrastive learning. Learning sentence embeddings with miCSE entails enforcing the syntactic consistency across augmented views for every single sentence, making contrastive self-supervised learning more sample efficient. As a result, the proposed appro… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(3 citation statements)
references
References 26 publications
0
3
0
Order By: Relevance
“…Refs. [4,[6][7][8][9][10][11][12][13][14][15] use contrastive learning. BERT-Flow and BERT-whitening [5,55] are post-processing models that apply flow-network and whitening to enhance BERT, respectively.…”
Section: Baseline and Previous Modelsmentioning
confidence: 99%
See 2 more Smart Citations
“…Refs. [4,[6][7][8][9][10][11][12][13][14][15] use contrastive learning. BERT-Flow and BERT-whitening [5,55] are post-processing models that apply flow-network and whitening to enhance BERT, respectively.…”
Section: Baseline and Previous Modelsmentioning
confidence: 99%
“…The best average result is in bold in the last column. †: [33], ‡: [9], ♠: [6], ♣: [10], : [56], : [4], ♥: [7], ♦: [8], : [13] , : [14], * : [15], SBERT-base-nli-v2: reproduced by ourselves, and the rest of the results are taken from Ref. [12].…”
Section: Reproducing Sbert-base-nli-v2 Modelmentioning
confidence: 99%
See 1 more Smart Citation