2018
DOI: 10.1007/978-3-030-01264-9_12
|View full text |Cite
|
Sign up to set email alerts
|

Weakly-Supervised Video Summarization Using Variational Encoder-Decoder and Web Prior

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
50
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 70 publications
(50 citation statements)
references
References 29 publications
0
50
0
Order By: Relevance
“…Our method outperforms all the baselines, including the supervised ranking-based methods [33,9]. and VESD [3]. We also implement a baseline where we train classifiers (CLA) with our hashtagged Instagram videos.…”
Section: Methodsmentioning
confidence: 98%
“…Our method outperforms all the baselines, including the supervised ranking-based methods [33,9]. and VESD [3]. We also implement a baseline where we train classifiers (CLA) with our hashtagged Instagram videos.…”
Section: Methodsmentioning
confidence: 98%
“…Our goal is analogous to the existing works on keyframe extraction from videos. The problem of keyframe extraction has been extensively studied in the context of video summarization [DKD98, CSJ15, ZCSG16, KVGUH18, DM18, CZDZ18] and scene change detection [MJC95, Sha95, HK17]. The goal of video summarization, however, is to find a compact set of images that can well represent as much video content as possible.…”
Section: Related Workmentioning
confidence: 99%
“…To some extent, our goal is analogous to the works of keyframe extraction from videos [DKD98, CSJ15, ZCSG16, KVGUH18, DM18, CZDZ18, MJC95, Sha95, HK17]. A key difference is that we need to extract the “keyframes” based on the path of walking, instead of video content.…”
Section: Introductionmentioning
confidence: 99%
“…Weakly supervised methods required a small number of annotations and could achieve great performance. [1] proposed a weakly supervised method that only required the topic label for a video. A variational autoencoder (VAE) model was trained by the massive edited videos with topic labels on the Internet to learn a better video representation.…”
Section: Related Work 21 Video Summarizationmentioning
confidence: 99%