2022
DOI: 10.1007/978-3-031-19833-5_29
|View full text |Cite
|
Sign up to set email alerts
|

Tip-Adapter: Training-Free Adaption of CLIP for Few-Shot Classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
28
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 82 publications
(39 citation statements)
references
References 31 publications
0
28
0
Order By: Relevance
“…Fine‐tuning the CLIP model or using advanced language‐image models ( e.g. , [LLXH22, LSG * 21, ZZF * 22]) are promising. Obtaining enough chart‐description pairs for training is also necessary, even the process is time‐consuming and resource‐intensive.…”
Section: Discussionmentioning
confidence: 99%
“…Fine‐tuning the CLIP model or using advanced language‐image models ( e.g. , [LLXH22, LSG * 21, ZZF * 22]) are promising. Obtaining enough chart‐description pairs for training is also necessary, even the process is time‐consuming and resource‐intensive.…”
Section: Discussionmentioning
confidence: 99%
“…Compared to CLIP-based few-shot adaption methods, including linear-probe CLIP, 4 CoOp, 5 WiSE-FT, 48 and Tip-Adapter-F, 66 our TEG combines the strengths of linear-probe CLIP for domain transfer and the guidance information from embedded text features, utilizing the proposed text-guided classifier to achieve superior performance in theme recognition. In addition, the auxiliary classification loss also helps the learning of visual concepts and stabilizes the network training.…”
Section: Methodsmentioning
confidence: 99%
“…A contrastive learning‐based image and NLP model developed by OpenAI. The Clip model combines unsupervised learning methods on large‐scale image and text corpora, aiming to embed text and images into the same space, allowing for similarity comparisons across different modalities. Tip‐Adapter (Zhang, Wei, et al., 2022 ). A lightweight and scalable model adaptation technology based on the Clip model, which enables rapid adaptation to new tasks or domains without the need to retrain the entire model.…”
Section: Methodsmentioning
confidence: 99%