2019
DOI: 10.1007/s00371-019-01666-x
|View full text |Cite|
|
Sign up to set email alerts
|

DeepLight: light source estimation for augmented reality using deep learning

Abstract: This paper presents a novel method for illumination estimation from RGB-D images. The main focus of the proposed method is to enhance visual coherence in augmented reality applications by providing accurate and temporally coherent estimates of real illumination. For this purpose, we designed and trained a deep neural network which calculates a dominant light direction from a single RGB-D image. Additionally, we propose a novel method for real-time outlier detection to achieve temporally coherent estimates. Our… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
15
0
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 36 publications
(16 citation statements)
references
References 30 publications
(43 reference statements)
0
15
0
1
Order By: Relevance
“…They integrated their network into an AR application and achieved a mean angular error of approximately 28 • and an inference time of approximately 380 ms while running on a central processing unit. The rendering equation (Section 1) suggests all required information to be present in an RGB image to reconstruct the illumination origin in a scene, however, given the unsuccessful experiment of Kán and Kaufmann (2019) using only RGB information, we begin with a simple scenario to investigate if RGB images provide enough information to reconstruct the dominant light direction. Considering the different dataset sizes and training strategies of Kán and Kaufmann (2019) and Marques et al (2018), a DNN with model weights pre-trained on a large dataset as initial values for training and a large number of training examples appears to be a reasonable basis for a regression approach that aims to reconstruct the dominant light direction from RGB images.…”
Section: State Of the Artmentioning
confidence: 99%
“…They integrated their network into an AR application and achieved a mean angular error of approximately 28 • and an inference time of approximately 380 ms while running on a central processing unit. The rendering equation (Section 1) suggests all required information to be present in an RGB image to reconstruct the illumination origin in a scene, however, given the unsuccessful experiment of Kán and Kaufmann (2019) using only RGB information, we begin with a simple scenario to investigate if RGB images provide enough information to reconstruct the dominant light direction. Considering the different dataset sizes and training strategies of Kán and Kaufmann (2019) and Marques et al (2018), a DNN with model weights pre-trained on a large dataset as initial values for training and a large number of training examples appears to be a reasonable basis for a regression approach that aims to reconstruct the dominant light direction from RGB images.…”
Section: State Of the Artmentioning
confidence: 99%
“…With 100 possible point light sources available in the virtual scene, their network achieved a top 1 accuracy of approximately 82% in classifying the correct point light source. Kán and Kaufmann (2019) proposed a regression model to reconstruct continuous azimuth and elevation angle values indicating the dominant light direction in real scenes using RGB-D input images for their ResNet. They also conducted experiments using the RGB com-ponents of their RGB-D dataset; however, their network failed when it relied solely on these three colour channels.…”
Section: State Of the Artmentioning
confidence: 99%
“…In our proposed approach, we aim to reconstruct the azimuth and elevation angles (φ , θ ) of the dominant light direction from RGB input images. This is similar to the approach of Kán and Kaufmann (2019); however, our approach focuses on light direction reconstruction using only RGB information and explicitly omits the depth information. We assume that the network is capable of reconstructing the light direction without additional depth information.…”
Section: Light Direction Inferencementioning
confidence: 99%
See 1 more Smart Citation
“…Pemanfaatannya dapat dirasakan dalam bidang pendidikan tinggi (Nguyen, et al, 2018), untuk pembelajaran keahlian bedah syaraf (Si, et al, 2019), perancangan brosur interaktif (Rumajar, et al, 2015), sarana promosi produk tertentu (Tijono, et al, 2015) dan banyak aplikasi lainnya. Dalam riset lain, penggabungan AR dan deep learning digunakan untuk mengestimasi pencahayaan benda (object illumination) pada lingkungan gambar nyata (Kán & Kafumann, 2019). Dalam penelitian tersebut, arah sumber cahaya disimulasi dengan deep learning sehingga menghasilkan arah bayangan yang tepat.…”
Section: Aplikasi Arunclassified