2020
DOI: 10.48550/arxiv.2006.06500
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Rethinking the Truly Unsupervised Image-to-Image Translation

Abstract: Figure 1. Our model is able to conduct image-to-image translation without any supervision. The output images are generated with the source image and the the average style code of each estimated domain. The breed of the output cat changes according to the domain while preserving the pose of the source image.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
25
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(26 citation statements)
references
References 37 publications
1
25
0
Order By: Relevance
“…For each task, we used a suitable baseline architecture, but replaced their content losses with our F/LSeSim loss. In addition, we are only interested in scenarios where scene structure is preserved during the translation [23,56,57], rather investigating translations incorporating shape modification [9,10,37,28,4].…”
Section: Methodsmentioning
confidence: 99%
“…For each task, we used a suitable baseline architecture, but replaced their content losses with our F/LSeSim loss. In addition, we are only interested in scenarios where scene structure is preserved during the translation [23,56,57], rather investigating translations incorporating shape modification [9,10,37,28,4].…”
Section: Methodsmentioning
confidence: 99%
“…Baselines. We use TUNIT [2] and SwapAE [31] as our unsupervised baselines. In the case of TUNIT, the number of clusters must be specified in advance, but it is difficult to know the optimal number for each dataset.…”
Section: Methodsmentioning
confidence: 99%
“…While successful, these methods rely on a vast quantity of domain labels, which often becomes a serious bottleneck. To reduce such appetite for labels, more recent studies propose fully unsupervised methods that leverage pseudolabels acquired by the image clustering methods [3,2]. However these methods easily yield unintended translation results if the clustering algorithms fail to produce consistent clusters.…”
Section: Multi-domain Image-to-image Translationmentioning
confidence: 99%
See 1 more Smart Citation
“…However, the above scenario [111], [161] still assumes access to the domain labels of the training images. Some recent work aims to reduce the need for such supervision by using few [209] or even no [9] domain labels. Very recently, some works [15], [106], [147] are able to achieve image translation even when each domain only has a single image, inspired by recent advances that can train GANs on a single image [169].…”
Section: Unsupervised Image Translationmentioning
confidence: 99%