2020
DOI: 10.1007/978-3-030-66096-3_27
|View full text |Cite
|
Sign up to set email alerts
|

Learning Temporally Invariant and Localizable Features via Data Augmentation for Video Recognition

Abstract: Data augmentation has recently emerged as an essential component of modern training recipes for visual recognition tasks. However, data augmentation for video recognition has been rarely explored despite its effectiveness. Few existing augmentation recipes for video recognition naively extend the image augmentation methods by applying the same operations to the whole video frames. Our main idea is that the magnitude of augmentation operations for each frame needs to be changed over time to capture the real-wor… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4

Citation Types

0
19
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 13 publications
(19 citation statements)
references
References 75 publications
(197 reference statements)
0
19
0
Order By: Relevance
“…We propose to linearly change the bounding pox positions for cut-and-paste algorithms, such as CutOut [11], CutMix [65], and CutMixUp [64], and their extensions, as well as the mixing ratio in MixUp [66] and CutMixUp [64] extensions; 4. The recognition results of the aforementioned techniques on the UCF-101 [55] and the HMDB-51 [38] datasets either maintain competitive or exceed performance achieved by the previous work [34].…”
Section: Introductionmentioning
confidence: 73%
See 4 more Smart Citations
“…We propose to linearly change the bounding pox positions for cut-and-paste algorithms, such as CutOut [11], CutMix [65], and CutMixUp [64], and their extensions, as well as the mixing ratio in MixUp [66] and CutMixUp [64] extensions; 4. The recognition results of the aforementioned techniques on the UCF-101 [55] and the HMDB-51 [38] datasets either maintain competitive or exceed performance achieved by the previous work [34].…”
Section: Introductionmentioning
confidence: 73%
“…In this paper, we expand on the previous work [34]. We argue that some of the proposed techniques can be extended even further to fully utilise the time domain and achieve a deeper level of temporal perturbations, which results in more accurate and robust classifiers.…”
Section: Introductionmentioning
confidence: 93%
See 3 more Smart Citations