2022
DOI: 10.1109/tpami.2021.3126682
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Moments in Time: Learning and Interpreting Models for Multi-Action Video Understanding

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
33
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 29 publications
(33 citation statements)
references
References 24 publications
0
33
0
Order By: Relevance
“…Dataset. We evaluate the proposed DEAR method on three commonly used real-world video action datasets, including UCF-101 [55], HMDB-51 [31], and MiT-v2 [39]. All models are trained on UCF-101 training split.…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations
“…Dataset. We evaluate the proposed DEAR method on three commonly used real-world video action datasets, including UCF-101 [55], HMDB-51 [31], and MiT-v2 [39]. All models are trained on UCF-101 training split.…”
Section: Methodsmentioning
confidence: 99%
“…Evaluation Protocol. To evaluate the classification per- [31] and MiT-v2 [39], respectively. For Open maF1 scores, both the mean and standard deviation of 10 random trials of unknown class selection are reported.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…More recent datasets consider videos at an atomic level, with fine-grained temporal annotations from short snippets of longer videos [25,49,84]. In particular, Multi-Moments in Time [50] provides 2M action labels for 1M short clips of 3s, classified into 313 annotated action classes. Something-Something [24] collects more than 100k videos annotated with 147 classes of daily human-object interactions.…”
Section: Related Workmentioning
confidence: 99%