2023
DOI: 10.1111/cgf.14911
|View full text |Cite
|
Sign up to set email alerts
|

Factored Neural Representation for Scene Understanding

Abstract: A long‐standing goal in scene understanding is to obtain interpretable and editable representations that can be directly constructed from a raw monocular RGB‐D video, without requiring specialized hardware setup or priors. The problem is significantly more challenging in the presence of multiple moving and/or deforming objects. Traditional methods have approached the setup with a mix of simplifications, scene priors, pretrained templates, or known deformation models. The advent of neural representations, espec… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
0
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 98 publications
(138 reference statements)
0
0
0
Order By: Relevance
“…BANMO further decomposes the object motion into articulated and non‐rigid motion, from which both RPD and TotalRecon benefit. FactoredNeRF [WM23] assumes the root motion and segmentation masks of objects are provided at keyframes, enabling the decomposition of the scene into the background and multiple moving objects. This factored representation allows for object manipulations including removing an object or changing an object's trajectory.…”
Section: State‐of‐the‐art Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…BANMO further decomposes the object motion into articulated and non‐rigid motion, from which both RPD and TotalRecon benefit. FactoredNeRF [WM23] assumes the root motion and segmentation masks of objects are provided at keyframes, enabling the decomposition of the scene into the background and multiple moving objects. This factored representation allows for object manipulations including removing an object or changing an object's trajectory.…”
Section: State‐of‐the‐art Methodsmentioning
confidence: 99%
“…3.2. Once the per‐object model is learned, individual objects can be moved, removed, duplicated, or their trajectories can be changed [WM23]. For scenes with many objects, this might be impractical as one NeRF per object needs to be learned, and controlling general properties of the scene, such as global illumination, is more challenging.…”
Section: State‐of‐the‐art Methodsmentioning
confidence: 99%