2019 IEEE/CVF International Conference on Computer Vision (ICCV) 2019
DOI: 10.1109/iccv.2019.00811
|View full text |Cite
|
Sign up to set email alerts
|

AWSD: Adaptive Weighted Spatiotemporal Distillation for Video Representation

Abstract: We propose an Adaptive Weighted Spatiotemporal Distillation (AWSD) technique for video representation by encoding the appearance and dynamics of the videos into a single RGB image map. This is obtained by adaptively dividing the videos into small segments and comparing two consecutive segments. This allows using pre-trained models on still images for video classification while successfully capturing the spatiotemporal variations in the videos. The adaptive segment selection enables effective encoding of the es… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 32 publications
0
3
0
Order By: Relevance
“…A thorough review of the field is provided in Gou et al (2021). Knowledge Distillation has been employed on various computer vision problems, i.e., image classification (Yalniz et al, 2019;Touvron et al, 2020;Xie et al, 2020), object detection (Li et al, 2017;Shmelkov et al, 2017;Deng et al, 2019), metric learning (Park et al, 2019;Peng et al, 2019), action recognition (Garcia et al, 2018;Thoker & Gall, 2019;Stroud et al, 2020), video classification (Zhang & Peng, 2018;Bhardwaj et al, 2019), video captioning (Pan et al, 2020;Zhang et al, 2020), and representation learning (Tavakolian et al, 2019;Piergiovanni et al, 2020).…”
Section: Knowledge Distillationmentioning
confidence: 99%
“…A thorough review of the field is provided in Gou et al (2021). Knowledge Distillation has been employed on various computer vision problems, i.e., image classification (Yalniz et al, 2019;Touvron et al, 2020;Xie et al, 2020), object detection (Li et al, 2017;Shmelkov et al, 2017;Deng et al, 2019), metric learning (Park et al, 2019;Peng et al, 2019), action recognition (Garcia et al, 2018;Thoker & Gall, 2019;Stroud et al, 2020), video classification (Zhang & Peng, 2018;Bhardwaj et al, 2019), video captioning (Pan et al, 2020;Zhang et al, 2020), and representation learning (Tavakolian et al, 2019;Piergiovanni et al, 2020).…”
Section: Knowledge Distillationmentioning
confidence: 99%
“…A thorough review of the field is provided in (Gou et al, 2021). Knowledge Distillation has been employed on various computer vision problems, i.e., image classification (Yalniz et al, 2019;Touvron et al, 2020;Xie et al, 2020), object detection (Li et al, 2017;Shmelkov et al, 2017;Deng et al, 2019), metric learning (Park et al, 2019;Peng et al, 2019), action recognition (Garcia et al, 2018;Thoker and Gall, 2019;Stroud et al, 2020), video classification (Zhang and Peng, 2018;Bhardwaj et al, 2019), video captioning (Pan et al, 2020;Zhang et al, 2020), and representation learning (Tavakolian et al, 2019;Piergiovanni et al, 2020).…”
Section: Knowledge Distillationmentioning
confidence: 99%
“…This work suggests that class probabilities, as "dark knowledge", are very useful to retain the performance of original network, and thus, light-weight substitute model could be trained to distill this knowledge. This approach is very useful and has been justified to solve a variety of complex application problems, such as pose estimation [37,46,33], lane detection [17], real-time streaming [31], object detection [6], video representation [41,10,11], and so forth. Furthermore, this approach is able to boost the performance of deep neural network with improvement on efficiency [35] and accuracy [25].…”
Section: Related Workmentioning
confidence: 99%