2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.01022
|View full text |Cite
|
Sign up to set email alerts
|

Out-Of-Distribution Detection for Generalized Zero-Shot Action Recognition

Abstract: Generalized zero-shot action recognition is a challenging problem, where the task is to recognize new action categories that are unavailable during the training stage, in addition to the seen action categories. Existing approaches suffer from the inherent bias of the learned classifier towards the seen action categories. As a consequence, unseen category samples are incorrectly classified as belonging to one of the seen action categories. In this paper, we set out to tackle this issue by arguing for a separate… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

2
99
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 116 publications
(101 citation statements)
references
References 27 publications
2
99
0
Order By: Relevance
“…We use = 2048-dimension ResNet101 [8] features for images in AWA1 and CUB while manually defined attributes of dimension = 85 and = 312 for AWA1 and CUB, respectively. For HMDB51 and UCF101, we use the I3D [2] video feature of = 8196 dimension and = 300-dim word2vec as semantic prototypes provided by [16]. For AWA1 and CUB we use the new seen/unseen split proposed by [34] while for HMDB1 and UCF101 we use the seen/unseen splits provided by [16] where we experiment on 30 splits and report the average performance.…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations
“…We use = 2048-dimension ResNet101 [8] features for images in AWA1 and CUB while manually defined attributes of dimension = 85 and = 312 for AWA1 and CUB, respectively. For HMDB51 and UCF101, we use the I3D [2] video feature of = 8196 dimension and = 300-dim word2vec as semantic prototypes provided by [16]. For AWA1 and CUB we use the new seen/unseen split proposed by [34] while for HMDB1 and UCF101 we use the seen/unseen splits provided by [16] where we experiment on 30 splits and report the average performance.…”
Section: Methodsmentioning
confidence: 99%
“…For HMDB51 and UCF101, we use the I3D [2] video feature of = 8196 dimension and = 300-dim word2vec as semantic prototypes provided by [16]. For AWA1 and CUB we use the new seen/unseen split proposed by [34] while for HMDB1 and UCF101 we use the seen/unseen splits provided by [16] where we experiment on 30 splits and report the average performance. We use the evaluation criterion proposed by [34] where we report the average top 1 accuracy on the seen ( ) and the unseen ( ) classes and the harmonic mean ( ) of and .…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Recently, a popular approach of zero-shot classification is generating synthesized features for unseen categories. For example, the method in [44] first generated features using word embeddings and random vectors, which was further improved by later works [7,22,28,40,45]. These zero-shot classification methods generated image features without involving contextual information.…”
Section: Related Workmentioning
confidence: 99%