2021
DOI: 10.1016/j.adhoc.2020.102380
|View full text |Cite
|
Sign up to set email alerts
|

VDARN: Video Disentangling Attentive Relation Network for Few-Shot and Zero-Shot Action Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
17
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 15 publications
(17 citation statements)
references
References 31 publications
0
17
0
Order By: Relevance
“…The former assumes that only the labeled videos from the seen categories are available during training while the latter can use the unlabeled data of the unseen categories for model training. Specifically, in this work, we focus on inductive ZSAR [12], [15], [26], [42] and do not discuss the transductive approach [9], [32].…”
Section: Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…The former assumes that only the labeled videos from the seen categories are available during training while the latter can use the unlabeled data of the unseen categories for model training. Specifically, in this work, we focus on inductive ZSAR [12], [15], [26], [42] and do not discuss the transductive approach [9], [32].…”
Section: Methodsmentioning
confidence: 99%
“…For example, Jain et al [14] extract objects from action videos by a pretrained object classifier and embed object names to form a visual representation. In [15], the object information and pose information are simultaneously extracted to form visual feature. However, both of them utilize the pretrained object classifier at test time, which is questionable as illustrated in the introduction.…”
Section: A Zero-shot Action Recognitionmentioning
confidence: 99%
See 3 more Smart Citations