2021
DOI: 10.48550/arxiv.2102.13090
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

IBRNet: Learning Multi-View Image-Based Rendering

Abstract: We present a method that synthesizes novel views of complex scenes by interpolating a sparse set of nearby views. The core of our method is a network architecture that includes a multilayer perceptron and a ray transformer that estimates radiance and volume density at continuous 5D locations (3D spatial locations and 2D viewing directions), drawing appearance information on the fly from multiple source views. By drawing on source views at render time, our method hearkens back to classic work on image-based ren… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
27
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(27 citation statements)
references
References 56 publications
0
27
0
Order By: Relevance
“…We demonstrate this by comparison to several state-of-the-art methods. Specifically, we evaluate the volumetric representation of NeRF [59], a meshbased representation similar to SVS [8], the neural signed distance function-based representations of IDR [4] and NLR [5], and the image-based rendering of IBRNet [28]. For SVS [8] we use our own simplified implementation and denote it SVS*.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…We demonstrate this by comparison to several state-of-the-art methods. Specifically, we evaluate the volumetric representation of NeRF [59], a meshbased representation similar to SVS [8], the neural signed distance function-based representations of IDR [4] and NLR [5], and the image-based rendering of IBRNet [28]. For SVS [8] we use our own simplified implementation and denote it SVS*.…”
Section: Methodsmentioning
confidence: 99%
“…Recent IBR techniques leverage neural networks to learn the required blending weights [19][20][21][22][23][24]. These neural IBR methods either use proxy geometry, for example obtained by SfM or MVS [25,26], together with on-surface feature aggregation [7,8] or use learned pixel aggregation functions [27,28] for geometry-free image-based view synthesis. Our approach is closely related to the geometry-assisted and feature-interpolating view synthesis techniques.…”
Section: Related Workmentioning
confidence: 99%
“…Baselines. Among the recent generalizable NeRF methods [36,52,47], we compare with Pixel-NeRF [52] and PVA [36] which focus on very sparse (up to 3 or 4) input views. we reimplement [36] since it is not open-sourced.…”
Section: Comparison With Generalizable Nerf Methodsmentioning
confidence: 99%
“…Despite the promising results, these general NeRF [19,53] and human-specific NeRF [13,32,33,35,50] methods must be optimized for each new video separately, and generalize poorly on unseen scenarios. Generalizable NeRFs [36,47,52] try to avoid the expensive per-scene optimization by imageconditioning using pixel-aligned features. However, directly extending such methods to model complex and dynamic 3D humans is not straightforward when available observations are highly sparse.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation