2020
DOI: 10.48550/arxiv.2005.13402
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

AVGZSLNet: Audio-Visual Generalized Zero-Shot Learning by Reconstructing Label Features from Multi-Modal Embeddings

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 0 publications
0
1
0
Order By: Relevance
“…As an example, a new multi-label ZSL (MZSL) framework using JLRE (Joint Latent Ranking Embedding) has been proposed in [158]. The relatedness score of various action labels is measured for the test video clips in the semantic embedding and joint latent visual spaces.In addition, a multimodal framework using audio, video, and text has been introduced in [160], [161].…”
Section: Applicationsmentioning
confidence: 99%
“…As an example, a new multi-label ZSL (MZSL) framework using JLRE (Joint Latent Ranking Embedding) has been proposed in [158]. The relatedness score of various action labels is measured for the test video clips in the semantic embedding and joint latent visual spaces.In addition, a multimodal framework using audio, video, and text has been introduced in [160], [161].…”
Section: Applicationsmentioning
confidence: 99%