2017
DOI: 10.48550/arxiv.1709.00663
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Generative Model For Zero Shot Learning Using Conditional Variational Autoencoders

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
24
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(24 citation statements)
references
References 0 publications
0
24
0
Order By: Relevance
“…For AWA, the preferred range is k ∈ [7,17] and the most preferred k is around 12. For CUB, the preferred range is k ∈ [4,19] and the most preferred k is around 16. Secondly, we evaluate the training time.…”
Section: Results and Analysismentioning
confidence: 99%
“…For AWA, the preferred range is k ∈ [7,17] and the most preferred k is around 12. For CUB, the preferred range is k ∈ [4,19] and the most preferred k is around 16. Secondly, we evaluate the training time.…”
Section: Results and Analysismentioning
confidence: 99%
“…Multi-label Zero-shot Learning is a less explored problem compared to single-label zero-shot learning [10,20,24,26,32]. Fu et al [7] enumerate all possible combinations of labels and thus transform the problem into single-label zero-shot learning.…”
Section: Related Workmentioning
confidence: 99%
“…Without the yellow nodes from ImageNet[23], it would be hard to predict the unseen class Sheep since other seen classes are not semantically related to it. However, with the knowledge that the current image has a high probability on the ImageNet class Bighorn Sheep and the word embeddings of both Sheep and Bighorn Sheep, the model can better predict the unseen class by referring to the ImageNet classes.Although zero-shot learning for single-label classification has been well studied[10,20,24,26,32,33], multi-label zero-shot learning (ML-ZSL)[7,14,18,35] is a less explored area. An image is associated with potentially multiple seen and unseen classes, but only labels of seen classes are provided during training.…”
mentioning
confidence: 99%
“…At test time, the label for an unseen-class test input is the class that maximizes the VAE lower bound. Afterwards, Mishra et al (2017) trained a conditional VAE (Sohn et al, 2015) to learn the underlying probability distribution of the image features conditioned on the class embedding vector. Similarly, Arora et al (2018) proposed a method able to generate semantically rich CNN feature distributions via the conditional VAE with discriminator-driven feedback mechanism improving the reconstruction capability.…”
Section: Extension Matchingmentioning
confidence: 99%