2022 International Conference on 3D Vision (3DV) 2022
DOI: 10.1109/3dv57658.2022.00046
|View full text |Cite
|
Sign up to set email alerts
|

3inGAN: Learning a 3D Generative Model from Images of a Self-similar Scene

Abstract: Figure 1. We present HOLODIFFUSION as the first 3D-aware generative diffusion model that produces 3D-consistent images and is trained with only posed image supervision. Here we show a few different samples generated from models trained on different classes of the CO3D dataset [50].

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2

Relationship

1
4

Authors

Journals

citations
Cited by 6 publications
(1 citation statement)
references
References 99 publications
0
1
0
Order By: Relevance
“…Neural Radiance Field (NeRF) [MST*20], a new volumetric neural representation, provided a breakthrough in terms of producing highly photorealistic (static) representation, simultaneously capturing geometry and appearance from only a set of posed images. A substantial body of work has rapidly emerged to extend the formulation to dynamic settings [LSZ*22,DZY*21,PCPMMN21,LNSW21,TTG*21, XHKK21,GSKH21], work with localized representations for real‐time inference [LGZL*20, RPLG21, YLT*21,LSS*21,SSC22,KRWM22b,FYW*22,WZL*22], support fast training [DLZR22, SSC22, KRWM22b, FYW*22, LCM*22], and investigate applications in the context of generative models [KRWM22a]. However, such representations often lack interpretability, require multiview input, fail to provide scene understanding, and do not provide object‐level factorization or enable object‐level scene manipulation.…”
Section: Introductionmentioning
confidence: 99%
“…Neural Radiance Field (NeRF) [MST*20], a new volumetric neural representation, provided a breakthrough in terms of producing highly photorealistic (static) representation, simultaneously capturing geometry and appearance from only a set of posed images. A substantial body of work has rapidly emerged to extend the formulation to dynamic settings [LSZ*22,DZY*21,PCPMMN21,LNSW21,TTG*21, XHKK21,GSKH21], work with localized representations for real‐time inference [LGZL*20, RPLG21, YLT*21,LSS*21,SSC22,KRWM22b,FYW*22,WZL*22], support fast training [DLZR22, SSC22, KRWM22b, FYW*22, LCM*22], and investigate applications in the context of generative models [KRWM22a]. However, such representations often lack interpretability, require multiview input, fail to provide scene understanding, and do not provide object‐level factorization or enable object‐level scene manipulation.…”
Section: Introductionmentioning
confidence: 99%