2022
DOI: 10.48550/arxiv.2205.10183
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Prototypical Calibration for Few-shot Learning of Language Models

Abstract: In-context learning of GPT-like models has been recognized as fragile across different hand-crafted templates, and demonstration permutations. In this work, we propose prototypical calibration to adaptively learn a more robust decision boundary for zero-and few-shot classification, instead of greedy decoding. Concretely, our method first adopts Gaussian mixture distribution to estimate the prototypical clusters for all categories. Then we assign each cluster to the corresponding label by solving a weighted bip… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
7
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(8 citation statements)
references
References 10 publications
1
7
0
Order By: Relevance
“…In short summary, in-context models are mainly biased by shifting the label marginal, which is generally implicitly assumed in previous works (Zhao et al, 2021;Han et al, 2022;Fei et al, 2023). However, to the best of our knowledge, we are the first to verify this systematically.…”
Section: In-context Label Marginal Is Differentsupporting
confidence: 59%
See 3 more Smart Citations
“…In short summary, in-context models are mainly biased by shifting the label marginal, which is generally implicitly assumed in previous works (Zhao et al, 2021;Han et al, 2022;Fei et al, 2023). However, to the best of our knowledge, we are the first to verify this systematically.…”
Section: In-context Label Marginal Is Differentsupporting
confidence: 59%
“…The data label marginal q(y) is hard to estimate since we only have a few training examples in our setting. Previous works (Zhao et al, 2021;Min et al, 2021;Han et al, 2022;Fei et al, 2023) typically assume q(y) to be uniform, yielding the following classifier:…”
Section: Generative Calibrationmentioning
confidence: 99%
See 2 more Smart Citations
“…The use of autoencoders in NVDLMED [ 8 ] for segmentation makes it more computationally expensive. In comparison to these models [ 41 ], our proposed approach contains fewer parameters and is not much computationally expensive. Furthermore, our one-shot segmentation model took a very low inference time for each image, which makes it very suitable to test unseen images in clinical settings.…”
Section: Resultsmentioning
confidence: 99%