2022
DOI: 10.48550/arxiv.2201.10963
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Learning to Compose Diversified Prompts for Image Emotion Classification

Abstract: Contrastive Language-Image Pre-training (CLIP) represents the latest incarnation of pre-trained vision-language models. Although CLIP has recently shown its superior power on a wide range of downstream vision-language tasks like Visual Question Answering, it is still underexplored for Image Emotion Classification (IEC). Adapting CLIP to the IEC task has three significant challenges, tremendous training objective gap between pretraining and IEC, shared suboptimal and invariant prompts for all instances. In this… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 4 publications
0
2
0
Order By: Relevance
“…Yi et al (2022) developed a contextual information and commonsense-based prompt learning model for conversational sentiment analysis, demonstrating superior performance over state-of-the-art models. Deng et al (2022b) also introduced a prompt tuning method that mimicked the pre-training objective of contrastive language-image pre-training (CLIP). It thus could leverage the rich image and text semantics for image emotion classification.…”
Section: Prompt Learningmentioning
confidence: 99%
“…Yi et al (2022) developed a contextual information and commonsense-based prompt learning model for conversational sentiment analysis, demonstrating superior performance over state-of-the-art models. Deng et al (2022b) also introduced a prompt tuning method that mimicked the pre-training objective of contrastive language-image pre-training (CLIP). It thus could leverage the rich image and text semantics for image emotion classification.…”
Section: Prompt Learningmentioning
confidence: 99%
“…To overcome the limitations of fixed prompts, many recent learned prompt works [19,60] propose to incorporate dataset-specific context information by using learnable prompt context vectors that can be catered to work best for each particular class. These learned prompts are biased towards seen classes.…”
Section: Prompt Learningmentioning
confidence: 99%