2018 International Conference on 3D Vision (3DV) 2018
DOI: 10.1109/3dv.2018.00012
|View full text |Cite
|
Sign up to set email alerts
|

Keep it Unreal: Bridging the Realism Gap for 2.5D Recognition with Geometry Priors Only

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
22
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
4

Relationship

3
6

Authors

Journals

citations
Cited by 29 publications
(22 citation statements)
references
References 26 publications
0
22
0
Order By: Relevance
“…As the eye training dataset RGB images are in different sensor space compared to the IR eye cameras inside the HMD, the detector might have a greater performance if trained with images that match the noise and spectral characteristics of the captured images. To decrease the domain gap between training data and application images, generative adversarial networks (GAN) have been proposed which learn to map between synthetic and real images with Rad et al [Rad et al 2019], Mueller et al [2018], and Zakharov et al [Zakharov et al 2018]. This approach could be used in our solution but it is necessary to have a different dataset to train the GAN network to learn the mapping between RBG and infrared images.…”
Section: Discussion and Future Workmentioning
confidence: 99%
“…As the eye training dataset RGB images are in different sensor space compared to the IR eye cameras inside the HMD, the detector might have a greater performance if trained with images that match the noise and spectral characteristics of the captured images. To decrease the domain gap between training data and application images, generative adversarial networks (GAN) have been proposed which learn to map between synthetic and real images with Rad et al [Rad et al 2019], Mueller et al [2018], and Zakharov et al [Zakharov et al 2018]. This approach could be used in our solution but it is necessary to have a different dataset to train the GAN network to learn the mapping between RBG and infrared images.…”
Section: Discussion and Future Workmentioning
confidence: 99%
“…The model was trained on UOAIS-Sim dataset following the standard schedule in [45] for 90, 000 iterations with SGD [46] using the learning rate of 0.00125. We applied color [47], depth [48], and crop augmentation. Training took about 8h on a single Tesla A100, and the inference took 0.13 s per image (1,000 iterations) on a Titan XP.…”
Section: Hierarchical Occlusion Modelingmentioning
confidence: 99%
“…The answer for this case is domain randomization. Domain randomization is a popular approach [49,21,55,36,47,40] that aims to randomize parts of the domain that we do not want our algorithm to be sensitive to. For example, [49] and [40] trained complex recognition methods by means of adding variability to the input render data, i.e., different illumination conditions, texture changes, scene decomposition, etc.…”
Section: Domain Randomizationmentioning
confidence: 99%
“…This sort of parameterization allows to learn features that are invariant to the particular properties of the domain. The authors of [55] used a sophisticated depth augmentation pipeline trying to cover possible artifacts of the common commodity depth sensors. It was then used to train a network removing these artifacts from the input and generating a clean, synthetically-looking image.…”
Section: Domain Randomizationmentioning
confidence: 99%