2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022
DOI: 10.1109/cvpr52688.2022.00539
|View full text |Cite
|
Sign up to set email alerts
|

Mip-NeRF 360: Unbounded Anti-Aliased Neural Radiance Fields

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
204
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 529 publications
(277 citation statements)
references
References 46 publications
1
204
0
Order By: Relevance
“…Thus, our approach yields to the segmentation error and still cannot provide high-quality hair rendering with opacity details. Inspired by recent work [Barron et al 2022;, we plan to explore the explicit modeling of background and opacity hairs into our neural framework. Moreover, our rendering pipeline relies on the animated meshes and is thus limited by the tracking quality.…”
Section: Limitations and Discussionmentioning
confidence: 99%
“…Thus, our approach yields to the segmentation error and still cannot provide high-quality hair rendering with opacity details. Inspired by recent work [Barron et al 2022;, we plan to explore the explicit modeling of background and opacity hairs into our neural framework. Moreover, our rendering pipeline relies on the animated meshes and is thus limited by the tracking quality.…”
Section: Limitations and Discussionmentioning
confidence: 99%
“…These methods take several hours or over a day to train and are intended for building a NeRF as opposed to real-time pose estimation with a trained NeRF. NeRF has also been extended to large-scale [26], [27], [28] and unbounded scenes [29], [30], which has the potential to enable neural representations of large-scale scenes such as the ones typically encountered in robotics applications, from drone navigation to self-driving cars.…”
Section: Related Workmentioning
confidence: 99%
“…The coordinate MLP M θ with a hidden layer dimension h = 256 maps a 3D position p = (x, y, z) ∈ R 3 into M θ (p) = (ρ, f , s), where ρ stands for volumetric density, f for a feature vector, and s for semantic logits. Before feeding a point p = (x, y, z) into the scene network M θ we apply off-axis positional encoding [16] to ensure rich representational capacity for fitting high-frequency feature maps and mitigate axis-aligned artifacts caused by the standard positional encoding used in NeRFs.…”
Section: A Scene Networkmentioning
confidence: 99%