2019 IEEE/CVF International Conference on Computer Vision (ICCV) 2019
DOI: 10.1109/iccv.2019.00774
|View full text |Cite
|
Sign up to set email alerts
|

Transformable Bottleneck Networks

Abstract: We propose a novel approach to performing fine-grained 3D manipulation of image content via a convolutional neural network, which we call the Transformable Bottleneck Network (TBN). It applies given spatial transformations directly to a volumetric bottleneck within our encoderbottleneck-decoder architecture. Multi-view supervision encourages the network to learn to spatially disentangle the feature space within the bottleneck. The resulting spatial structure can be manipulated with arbitrary spatial transforma… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

2
70
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 71 publications
(72 citation statements)
references
References 32 publications
2
70
0
Order By: Relevance
“…After our initial submission, several new related works have been published by the scientific community. We include here a non exhaustive list of them; [36], [68], [55], [37], [38], [7].…”
Section: Related Workmentioning
confidence: 99%
“…After our initial submission, several new related works have been published by the scientific community. We include here a non exhaustive list of them; [36], [68], [55], [37], [38], [7].…”
Section: Related Workmentioning
confidence: 99%
“…Recent novel view synthesis methods approach the single-image setting using deep learning [Tatarchenko et al 2015;Zhou et al 2016]. Synthesizing novel views from a single image is inherently challenging and existing methods are often only applicable to specific scene types [Habtegebrial et al 2018;Nguyen-Phuoc et al 2019], 3D object models [Olszewski et al 2019;Park et al 2017;Rematas et al 2017;Yan et al 2016;Yang et al 2015], or domainspecific light field imagery [Srinivasan et al 2017]. Most relevant to our work are methods that estimate the scene geometry of the input image via depth [Cun et al 2019;], normal maps , or layered depth [Tulsiani et al 2018].…”
Section: Learning-based View Synthesis From a Single Imagementioning
confidence: 99%
“…Recent work leverages deep neural networks to learn a monocular image-toimage mapping between source and target view from data [29,52,8,62,40,51,59]. One line of work [29,52,8,39] directly generates image pixels. Given the difficulty of the task, direct image-to-image translation approaches struggle with preservation of local details and often produce blurry images.…”
Section: Related Workmentioning
confidence: 99%