2018
DOI: 10.48550/arxiv.1810.01325
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

FutureGAN: Anticipating the Future Frames of Video Sequences using Spatio-Temporal 3d Convolutions in Progressively Growing GANs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
24
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 16 publications
(24 citation statements)
references
References 0 publications
0
24
0
Order By: Relevance
“…Video prediction is a challenging task due to the complex task of modelling both the content and motion in videos. To this extent several studies have been carried out to perform video prediction using GAN-based training ( [114], [164], [2], [66], [104], [166], [108]). Mathieu et al [114] used multi-scale architecture for future frame prediction.…”
Section: Video Predictionmentioning
confidence: 99%
“…Video prediction is a challenging task due to the complex task of modelling both the content and motion in videos. To this extent several studies have been carried out to perform video prediction using GAN-based training ( [114], [164], [2], [66], [104], [166], [108]). Mathieu et al [114] used multi-scale architecture for future frame prediction.…”
Section: Video Predictionmentioning
confidence: 99%
“…Predicting the whereabouts of an emerging object inherits predicting the future environment itself. Predicting the future environment was addressed by predicting future frames [55,52,51,1,28,32,61] and future semantic segmentation [34,25,54,33,8,6]. These methods can only hallucinate new objects in the scene in a photorealistic way, but none of them explicitly predicts the structure where new objects can actually emerge.…”
Section: Related Workmentioning
confidence: 99%
“…In this paper (1) we propose a future localization framework in egocentric view by transferring the work by Makansi et al [36] from bird's-eye view to egocentric observations, where multimodality is even more difficult to capture. Thus, (2) we propose to compute a reachability prior as intermediate result, which serves as attention to prevent forgetting rare modes, and which can be used to efficiently propagate scene priors into the future taking into account the egomotion.…”
Section: Introductionmentioning
confidence: 99%
“…Video Generation. Many approaches use GANs [1,34,36] or adversarial loss during training for generating videos [1,21,23,25,31,36,37]. Vondrick et al [36] use a generator with two pathways to predict the foreground and background, and a discriminator to distinguish a video as real or fake.…”
Section: Related Workmentioning
confidence: 99%