2018
DOI: 10.1145/3272127.3275099
|View full text |Cite
|
Sign up to set email alerts
|

LookinGood

Abstract: Fig. 1. LookinGood leverages recent advances in real-time 3D performance capture and machine learning to re-render high quality novel viewpoints of a captured scene. A textured 3D reconstruction is first rendered to a novel viewpoint. Due to imperfections in geometry and low-resolution texture, the 2D rendered image contains artifacts and is low quality. Therefore we propose a deep learning technique that takes these images as input and generates more visually enhanced re-rendering. The system is specifically … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 95 publications
(5 citation statements)
references
References 69 publications
0
5
0
Order By: Relevance
“…While some approaches have shown convincing results for the facial area [Kim et al 2018a;Lombardi et al 2018], creating photo-real images of the entire human is still a challenge. Most of the methods, which target synthesizing entire humans, learn an image-to-image mapping from renderings of a skeleton [Chan et al 2019;Esser et al 2018;Pumarola et al 2018;Si et al 2018], depth map [Martin-Brualla et al 2018], dense mesh [Liu et al 2020b[Liu et al , 2019bWang et al 2018a] or joint position heatmaps [Aberman et al 2019], to real images. Among these approaches, the most related work [Liu et al 2020b] achieves better temporally-coherent dynamic textures by first learning fine scale details in texture space and then translating the rendered mesh with dynamic textures into realistic imagery.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…While some approaches have shown convincing results for the facial area [Kim et al 2018a;Lombardi et al 2018], creating photo-real images of the entire human is still a challenge. Most of the methods, which target synthesizing entire humans, learn an image-to-image mapping from renderings of a skeleton [Chan et al 2019;Esser et al 2018;Pumarola et al 2018;Si et al 2018], depth map [Martin-Brualla et al 2018], dense mesh [Liu et al 2020b[Liu et al , 2019bWang et al 2018a] or joint position heatmaps [Aberman et al 2019], to real images. Among these approaches, the most related work [Liu et al 2020b] achieves better temporally-coherent dynamic textures by first learning fine scale details in texture space and then translating the rendered mesh with dynamic textures into realistic imagery.…”
Section: Related Workmentioning
confidence: 99%
“…Different from differentiable rendering, neural rendering makes almost no assumptions about the physical model and uses neural networks to learn the rendering process from data to synthesize photo-realistic images. Some neural rendering methods [Aberman et al 2019;Chan et al 2019;Kim et al 2018b;Liu et al 2020bLiu et al , 2019bMa et al 2017Ma et al , 2018Martin-Brualla et al 2018;Pumarola et al 2018;Shysheya et al 2019;Siarohin et al 2018;Thies et al 2019;Yoon et al 2020] employ image-to-image translation networks [Isola et al 2017;Wang et al 2018a,b] to augment the quality of the rendering. However, most of these methods suffer from view and/or temporal inconsistency.…”
Section: Related Workmentioning
confidence: 99%
“…To circumvent both the limitation to low resolution images associated with adversarial learning, and to remove ambiguities due to flat 2d maps (e.g. 2d skeletons missing 3d surface orientation), methods have utilized paired images produced by first rendering coarse geometry [18,14,19,15,20] and then translating to the corresponding real-world counter-part.…”
Section: Image-to-image Translation With a Proxy Meshmentioning
confidence: 99%
“…The rendered geometry allows to condition the network with less ambiguous data: a UV map [19], a low resolution avatar with patterned clothing [14], a rendered 3d skeleton [15], or noisy captured albedo and normals [18,20]. All these methods first render the 3d geometry and thus require creating and manipulating this intermediate object in their pipeline.…”
Section: Image-to-image Translation With a Proxy Meshmentioning
confidence: 99%
See 1 more Smart Citation

Rig-space Neural Rendering

Borer,
Yuhang,
Wuelfroth
et al. 2020
Preprint