2020
DOI: 10.1016/j.imavis.2020.103886
|View full text |Cite
|
Sign up to set email alerts
|

GANILLA: Generative adversarial networks for image to illustration translation

Abstract: In this paper, we explore illustrations in children's books as a new domain in unpaired image-to-image translation. We show that although the current state-of-the-art image-to-image translation models successfully transfer either the style or the content, they fail to transfer both at the same time. We propose a new generator network to address this issue and show that the resulting network strikes a better balance between style and content.There are no well-defined or agreed-upon evaluation metrics for unpair… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
40
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 51 publications
(40 citation statements)
references
References 33 publications
0
40
0
Order By: Relevance
“…The animation images are also segmented manually according to the same semantic classes of the photo images. The previous models [1], [2], [20], [21], [36]- [38] were retrained with these datasets for fair comparisons. All images were resized to 256 × 256 for training.…”
Section: A Implementation Detailsmentioning
confidence: 99%
See 2 more Smart Citations
“…The animation images are also segmented manually according to the same semantic classes of the photo images. The previous models [1], [2], [20], [21], [36]- [38] were retrained with these datasets for fair comparisons. All images were resized to 256 × 256 for training.…”
Section: A Implementation Detailsmentioning
confidence: 99%
“…Fig. 7 compares the qualitative results of our method with the CycleGAN [1], DiscoGAN [21], DualGAN [20], MUNIT [36], CartoonGAN [2], U-GAT-IT [37], and GANILLA [38]. It can be seen by comparing the results object-by-object that our method produces the best results with the fine details of the target anime object styles.…”
Section: A Implementation Detailsmentioning
confidence: 99%
See 1 more Smart Citation
“…Chen et al [116] propose an adversarial gated networks, called Gated-GAN, to transfer multiple styles while using a single model based on three modalities: an encoder, a gated transformer, and a decoder. GANILLA [117] is a proposed novel framework with the ability to better balance between content and style. Style transfer is the process of rendering the content of an image with a specific style while preserving the content, as shown in Figure 6.…”
Section: Style Transfermentioning
confidence: 99%
“…Metrics of success of image-to-image translation usually evaluate the quality of generated images while using a limited number of test images or user studies. The evaluation of a limited number of test images must consider both style and content simultaneously, which is difficult to do [117]. In addition, user studies are based on human judgment, which is a subjective metric [1].…”
Section: Lack Of Evaluation Metricsmentioning
confidence: 99%