2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.00748
|View full text |Cite
|
Sign up to set email alerts
|

Semantic Pyramid for Image Generation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
33
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 63 publications
(33 citation statements)
references
References 14 publications
0
33
0
Order By: Relevance
“…Characterizing such features has been a key question in scene perception for decades (Davenport & Potter, 2004;Greene & Oliva, 2009;Kauffmann et al, 2015;Oliva & Schyns, 2000;Oliva & Torralba, 2001;Schyns & Oliva, 1994;Walther et al, 2011); however, the recent revolution in deep learning approaches from machine learning and computer vision have provided novel insights into the nature of features underlying visual categorization (Cichy et al, 2017;Eberhardt et al, 2016;Krizhevsky et al, 2012;Rezanejad et al, 2019;Zhou et al, 2017;for review, Serre, 2019). In particular, generative adversarial networks (GAN; Brock et al, 2018;Goodfellow et al, 2014;Karras et al, 2019;Shocher et al, 2020;Yang et al, 2019) offer a data-driven method for uncovering the complex features spaces necessary for generating artificial but highly realistic images. GANs are composed of two deep neural networks, a generative network and a discriminative network, which perform antagonistic tasks.…”
mentioning
confidence: 99%
See 1 more Smart Citation
“…Characterizing such features has been a key question in scene perception for decades (Davenport & Potter, 2004;Greene & Oliva, 2009;Kauffmann et al, 2015;Oliva & Schyns, 2000;Oliva & Torralba, 2001;Schyns & Oliva, 1994;Walther et al, 2011); however, the recent revolution in deep learning approaches from machine learning and computer vision have provided novel insights into the nature of features underlying visual categorization (Cichy et al, 2017;Eberhardt et al, 2016;Krizhevsky et al, 2012;Rezanejad et al, 2019;Zhou et al, 2017;for review, Serre, 2019). In particular, generative adversarial networks (GAN; Brock et al, 2018;Goodfellow et al, 2014;Karras et al, 2019;Shocher et al, 2020;Yang et al, 2019) offer a data-driven method for uncovering the complex features spaces necessary for generating artificial but highly realistic images. GANs are composed of two deep neural networks, a generative network and a discriminative network, which perform antagonistic tasks.…”
mentioning
confidence: 99%
“…After successful learning, the synthesized images created by the generative network are highly realistic even to human observers. Recent studies of scene specific GANs (Shocher et al, 2020;Yang et al, 8 2019) have demonstrated that latent spaces from trained generative networks show a meaningful organization such that adjacent vectors in latent space are highly similar to each other and their associated synthetic images represent continuously changing realistic scenes. These advances offer a unique possibility -a continuous space of natural scene images may be constructed by deliberately traversing the latent space of a GAN trained to synthesize scene images.…”
mentioning
confidence: 99%
“…Initially, linear normalization [17] was applied on the dataset as a pre-processing stage before applying the image registration processes. Then, the Gaussian image pyramid [18,19] was used for down sampling to reduce processing time and memory size.…”
Section: Methodsmentioning
confidence: 99%
“…The image pyramid [26] is a collection of images at different scales as multi-scale representations of images. The semantic pyramid [27] is built on the basis of the image pyramid, and the model is constructed as a semantic pyramidal generation: the low-level information contains fine features (texture details, etc. ), while the high-level information overlays high-level semantic information.…”
Section: Semantic Pyramid Networkmentioning
confidence: 99%
“…Moreover, the residual blocks are added to each layer of the pyramid network to address the problem of reduced accuracy caused by the increase in the depth of the neural network. Therefore, we present the residual pyramid encoder built on the semantic pyramid structure [27]. It encodes the damaged images into compact latent features and decodes these features back to the image.…”
Section: Residual Pyramid Encodermentioning
confidence: 99%