2018 24th International Conference on Pattern Recognition (ICPR) 2018
DOI: 10.1109/icpr.2018.8545464
|View full text |Cite
|
Sign up to set email alerts
|

In2I: Unsupervised Multi-Image-to-Image Translation Using Generative Adversarial Networks

Abstract: In unsupervised image-to-image translation, the goal is to learn the mapping between an input image and an output image using a set of unpaired training images. In this paper, we propose an extension of the unsupervised image-toimage translation problem to multiple input setting. Given a set of paired images from multiple modalities, a transformation is learned to translate the input into a specified domain. For this purpose, we introduce a Generative Adversarial Network (GAN) based framework along with a mult… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
19
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
2
2

Relationship

2
6

Authors

Journals

citations
Cited by 31 publications
(19 citation statements)
references
References 19 publications
0
19
0
Order By: Relevance
“…Furthermore, when we compare the two ICVL figures [ 11 , 12 ] with the two previous CAVE figures [ 8 , 9 ], respectively, we may notice that the identical method performs more effectively on the ICVL scenes than on the CAVE scenes. We think that the main reason for this is that since most of the CAVE images are taken indoors, only a few bright objects contain useful information, and the main dark background contains scarce data to train the model.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Furthermore, when we compare the two ICVL figures [ 11 , 12 ] with the two previous CAVE figures [ 8 , 9 ], respectively, we may notice that the identical method performs more effectively on the ICVL scenes than on the CAVE scenes. We think that the main reason for this is that since most of the CAVE images are taken indoors, only a few bright objects contain useful information, and the main dark background contains scarce data to train the model.…”
Section: Methodsmentioning
confidence: 99%
“…Moreover, most researchers believe that complex deep neural networks would improve the models’ performance, and the more complex the neural networks are, the better effect they will have. Therefore, their approaches have often adopted tens or hundreds of convolutional neural network (CNN) layers with hundreds of residual network layers, which leads to a long computing time and high power consumption [ 11 , 12 , 13 ]. According to our thorough investigations and experiments, normal 2D convolutional operations scarcely improve the spectrum reconstruction performance and make the generated images blurry.…”
Section: Introductionmentioning
confidence: 99%
“…The success of GANs in synthesizing realistic images has led researchers to explore the GAN framework for numerous applications such as data augmentation [41], zeroshot learning [81], image inpainting [71], image dehazing [67,75,69], text-to-image translation [66], image-to-image translation [22,78,43], texture synthesis [23], crowd-counting [55] and generating outdoor scenes from attributes [25]. Isola et al [22] proposed a general method for image-to-image translation using conditional adversar'ial networks.…”
Section: Generative Adversarial Networkmentioning
confidence: 99%
“…Another issue with translation methods is that in the past, researchers have mainly focused on translation between two domains in the RGB modality. However, recently several researchers have had success with near-infrared (NIR) to RGB translation [29,28,23,18,22,27]. Also, CNN based translation methods help translate IR images to RGB for facial recognition [31,25].…”
Section: Image Translation Networkmentioning
confidence: 99%