2019
DOI: 10.1145/3306346.3323042
|View full text |Cite
|
Sign up to set email alerts
|

Deep inverse rendering for high-resolution SVBRDF estimation from an arbitrary number of images

Abstract: Fig. 1. Visualizations under natural lighting of four captured 1k resolution SVBRDFs estimated using our deep inverse rendering framework. The leather material (left) is reconstructed from just 2 input photographs captured with a mobile phone camera and flash, while the other materials are recovered from 20 input photographs. In this paper we present a unified deep inverse rendering framework for estimating the spatially-varying appearance properties of a planar exemplar from an arbitrary number of input photo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
128
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 143 publications
(129 citation statements)
references
References 22 publications
(56 reference statements)
1
128
0
Order By: Relevance
“…We demonstrate that our GAN-based optimization framework produces high-quality SVBRDF reconstructions from a small number (3-7) images captured under flash illumination using hand-held mobile phones, and improves upon previous state-of-the-art methods [Deschaintre et al 2019;Gao et al 2019]. In particular, it produces cleaner, more realistic material maps that better reproduce the appearance of the captured material under both input and novel lighting.…”
Section: :2 • Yumentioning
confidence: 75%
See 4 more Smart Citations
“…We demonstrate that our GAN-based optimization framework produces high-quality SVBRDF reconstructions from a small number (3-7) images captured under flash illumination using hand-held mobile phones, and improves upon previous state-of-the-art methods [Deschaintre et al 2019;Gao et al 2019]. In particular, it produces cleaner, more realistic material maps that better reproduce the appearance of the captured material under both input and novel lighting.…”
Section: :2 • Yumentioning
confidence: 75%
“…As a result, the reconstructed material parameters may not accurately reproduce the measured appearance. In contrast, Gao et al [2019] propose using an optimization-based approach in conjunction with a learned material prior. Specifically, they train a fully-convolutional auto-encoder on a large material dataset and optimize in the latent space of this auto-encoder.…”
Section: :2 • Yumentioning
confidence: 99%
See 3 more Smart Citations