2018
DOI: 10.48550/arxiv.1804.03608
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Imagine This! Scripts to Compositions to Videos

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(6 citation statements)
references
References 0 publications
0
6
0
Order By: Relevance
“…Sequentially generating new data from the previous data is termed autoregressive. However, we consider some studies [243,244] autoregressive due to the sequential prediction of frames, similar to others, but without using GAN or VAE models. These models typically fuse the two domains, text and video, for learning joint embedding.…”
Section: Auto-regressive Modelsmentioning
confidence: 99%
See 2 more Smart Citations
“…Sequentially generating new data from the previous data is termed autoregressive. However, we consider some studies [243,244] autoregressive due to the sequential prediction of frames, similar to others, but without using GAN or VAE models. These models typically fuse the two domains, text and video, for learning joint embedding.…”
Section: Auto-regressive Modelsmentioning
confidence: 99%
“…In CRAFT [243], text-conditioned video creation is completed by a compositional retrieval task. Following the caption, the model sequentially predicts a temporal layout of objects and retrieves the Spatio-temporal entity segments from a video dataset, where the fused segments create the final video.…”
Section: Generationmentioning
confidence: 99%
See 1 more Smart Citation
“…Pre-processing CIFAR-10 [ Krizhevsky, 2009] 32 × 32 × 3 64 None CelebA 64 × 64 × 3 128 Centre-cropped, area downsampled ImageNet [Deng et al, 2009] 64 × 64 × 3 128 Area downsampled Flintstones [Gupta et al, 2018] Since the normal GAN has no encoder, it was not necessary to add additional hyper-parameters when adding the losses in this case. For instance, the losses for each component for GAN + adversarial Z are:…”
Section: Models Consideredmentioning
confidence: 99%
“…The analysis of comics and mangas images recently sparked the computer vision and document analysis communities interest [2]. The digital version of manga can be used by the researchers to propose new algorithms to provide services such as dynamic visualization of manga [3], adding colors [12], generating animations [8], creating new kinds of recommender systems [7], etc.…”
Section: Introductionmentioning
confidence: 99%