2021
DOI: 10.48550/arxiv.2103.10428
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Large Scale Image Completion via Co-Modulated Generative Adversarial Networks

Abstract: Numerous task-specific variants of conditional generative adversarial networks have been developed for image completion. Yet, a serious limitation remains that all existing algorithms tend to fail when handling large-scale missing regions. To overcome this challenge, we propose a generic new approach that bridges the gap between image-conditional and recent modulated unconditional generative architectures via co-modulation of both conditional and stochastic style representations. Also, due to the lack of good … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
43
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 24 publications
(51 citation statements)
references
References 70 publications
(85 reference statements)
0
43
0
Order By: Relevance
“…ProFill [52] utilizes a contextual attention module and progressively fills the hole with predicted high-confidence pixels. CoModGAN [56] proposes a co-modulation of both conditional and stochastic representations to fill high quality content. Lastly, we compare a two-view SfM based approach [57] -which we refer to as JointDP -by warping the source image with the jointly estimated relative pose and depth using dense correspondence.…”
Section: Methodsmentioning
confidence: 99%
“…ProFill [52] utilizes a contextual attention module and progressively fills the hole with predicted high-confidence pixels. CoModGAN [56] proposes a co-modulation of both conditional and stochastic representations to fill high quality content. Lastly, we compare a two-view SfM based approach [57] -which we refer to as JointDP -by warping the source image with the jointly estimated relative pose and depth using dense correspondence.…”
Section: Methodsmentioning
confidence: 99%
“…To overcome this issue, Zheng et al [48] and Zhao et al [46] propose a VAE-based network that trade-offs between diversity and reconstruction. Zhao et al [47], inspired by the StyleGAN2 [15] modulated convolutions, introduces a comodulation layer for the inpainting task in order to improve both diversity and reconstruction. Noteworthy, a new family of auto-regressive methods [27,34,42], which can handle irregular masks, has recently emerged as a powerful alternative for free-form image inpainting.…”
Section: Related Workmentioning
confidence: 99%
“…However, the enforcement of diversity deteriorates the image quality. Inspired by the recent modulation approach [53] for multimodal image inpainting, we propose a similar network architecture specifically for open-domain image editing. The difference is that our modulation layer does not use the features of the input image, which leads to better fidelity and diversity.…”
Section: Related Workmentioning
confidence: 99%
“…Note that we circumvent direct pixel supervision such as L1 loss [16] for the purpose of encouraging the generation diversity, as suggested in [53]. Some qualitative output results from our trained generator is visualized in Fig.…”
Section: Multimodal Image Editing As Pretrainingmentioning
confidence: 99%
See 1 more Smart Citation