2020
DOI: 10.1007/978-3-030-66415-2_42
|View full text |Cite
|
Sign up to set email alerts
|

Using Sentences as Semantic Representations in Large Scale Zero-Shot Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 7 publications
(3 citation statements)
references
References 5 publications
0
3
0
Order By: Relevance
“…Incorporating language in the zero/fewshot setting has been widely explored. Embedding language from class names or descriptions to obtain class "prototypes" is common in zero-shot learning, when no visual samples of the class are available [7,8,17,31,32]. Several works also aim to learn classes using their semantic attributes for better knowledge transfer [16,25,45].…”
Section: Related Workmentioning
confidence: 99%
“…Incorporating language in the zero/fewshot setting has been widely explored. Embedding language from class names or descriptions to obtain class "prototypes" is common in zero-shot learning, when no visual samples of the class are available [7,8,17,31,32]. Several works also aim to learn classes using their semantic attributes for better knowledge transfer [16,25,45].…”
Section: Related Workmentioning
confidence: 99%
“…Following [PG11], the authors of [LCPLB20] further suggested to address the problem of bulk tagging [OM13] -users attributing the exact same tags to numerous photos -by ensuring that a tuple of words (wi, wj) can only appear once for each user during training, thus preventing a single user from having a disproportionate weight on the final embedding. Also, [LCLBC20] suggested to exploit the sentence descriptions of WordNet concepts, in addition to the class name embedding, to produce semantic representations better reflecting visual relations. Any of these two proposals allow to reach an accuracy between 17.2 and 17.8 on the 500 test classes of the ImageNet ZSL benchmark with the linear model from the semantic to the visual space (Section 3.1), compared to 14.4 with semantic prototypes based on standard embeddings.…”
Section: Semantic Features For Large Scale Zslmentioning
confidence: 99%
“…Ablation III: Effect of sentence embeddings for semantic matching. Sentence embeddings have recently been shown to be beneficial for zero-shot recognition in the image domain [22]. Here, we investigate their potential in the video domain.…”
Section: Ablation Studies On Object-scene Compositionsmentioning
confidence: 99%