2019 IEEE Winter Conference on Applications of Computer Vision (WACV) 2019
DOI: 10.1109/wacv.2019.00173
|View full text |Cite
|
Sign up to set email alerts
|

Attentive and Adversarial Learning for Video Summarization

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
54
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 63 publications
(55 citation statements)
references
References 21 publications
0
54
0
Order By: Relevance
“…Experiments on three public datasets SumMe, TVSum and YouTube demonstrate the effectiveness of our proposed framework. In future work, we will continue to investigate this line of research by utilizing reinforcement learning algorithm (Fu et al, 2019), attention mechanism (Ji et al, 2019) and multi-stage learning (Huang et al, 2019) within the DTR-GAN framework to further improve generic video summarization.…”
Section: Resultsmentioning
confidence: 99%
“…Experiments on three public datasets SumMe, TVSum and YouTube demonstrate the effectiveness of our proposed framework. In future work, we will continue to investigate this line of research by utilizing reinforcement learning algorithm (Fu et al, 2019), attention mechanism (Ji et al, 2019) and multi-stage learning (Huang et al, 2019) within the DTR-GAN framework to further improve generic video summarization.…”
Section: Resultsmentioning
confidence: 99%
“…[13] formulates video summarization as a sequence-to-sequence learning problem and proposes an LSTMbased encoder-decoder network with an intermediate attention layer. In [9], the typical encoder-decoder seq2seq model is replaced by a special attention-based seq2seq model that defines and ranks the different fragments of the video, and is combined with a 3D-CNN classifier which judges whether a fragment is from a ground-truth or a generated summary. [8] introduces an architecture with memory augmented networks for global attention modeling, and tackles video summarization by estimating the temporal dependency across the entire video.…”
Section: Related Workmentioning
confidence: 99%
“…The contributions of our work are: i) the introduction of an attention mechanism in an unsupervised learning framework, whereas all previous attentionbased summarization methods ([7-9, 13]) were supervised; ii) the investigation of integrating attention into a variational auto-encoder for video summarization purposes; and iii) the use of attention to guide the generative adversarial training of the model, rather than using it to rank the video fragments as in [9].…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Hyperlapse techniques sample frames adaptively by searching the optimal configuration (e.g., shortest path in a graph or dynamic programming) in a representation space where different features are combined to represent frames or frame transitions. Although recent works achieved better results applying a large number of features to represent the data [31]- [33], it increases both the computation time and memory usage since it leads to a high-dimensional space in optimization problems. We address this representation problem using a sparse frame sampling approach as depicted in Fig.…”
Section: B Weighted Sparse Frame Samplingmentioning
confidence: 99%