2020
DOI: 10.1109/tii.2019.2940136
|View full text |Cite
|
Sign up to set email alerts
|

Assembling Convolution Neural Networks for Automatic Viewing Transformation

Abstract: Images taken under different camera poses are rotated or distorted, which leads to poor perception experiences. This paper proposes a new framework to automatically transform the images to the conformable view setting by assembling different convolution neural networks. Specifically, a referential 3D ground plane is firstly derived from the RGB image and a novel projection mapping algorithm is developed to achieve automatic viewing transformation. Extensive experimental results demonstrate that the proposed me… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 26 publications
(47 reference statements)
0
2
0
Order By: Relevance
“…Blind distortion correction is an ill-posed problem. Therefore, learning-based methods using only a single distorted image are being pursued [ 10 , 11 , 12 , 42 , 43 , 44 , 45 , 46 ]. Deep learning for correcting documents were proposed recently [ 12 , 44 , 45 , 46 ] which implements convolutional neural networks, encoder-decoders, and U-net-based architectures [ 47 ].…”
Section: Related Workmentioning
confidence: 99%
“…Blind distortion correction is an ill-posed problem. Therefore, learning-based methods using only a single distorted image are being pursued [ 10 , 11 , 12 , 42 , 43 , 44 , 45 , 46 ]. Deep learning for correcting documents were proposed recently [ 12 , 44 , 45 , 46 ] which implements convolutional neural networks, encoder-decoders, and U-net-based architectures [ 47 ].…”
Section: Related Workmentioning
confidence: 99%
“…Depth is one of the most common projection types since there exists an explicit geometric relationship between the image coordinates in the source and target images for projection source view target view IBR result from [1] our result using depth. Most previous methods [2,[12][13][14][15] predict the depth of the input image to forward warp pixels of the source image. However, several source pixels might be mapped to one target location, which in turn will result in holes for pixels at unassigned target locations.…”
Section: Introductionmentioning
confidence: 99%