The platform will undergo maintenance on Sep 14 at about 7:45 AM EST and will be unavailable for approximately 2 hours.
2019 IEEE/CVF International Conference on Computer Vision (ICCV) 2019
DOI: 10.1109/iccv.2019.01007
|View full text |Cite
|
Sign up to set email alerts
|

UprightNet: Geometry-Aware Camera Orientation Estimation From Single Images

Abstract: We introduce UprightNet, a learning-based approach for estimating 2DoF camera orientation from a single RGB image of an indoor scene. Unlike recent methods that leverage deep learning to perform black-box regression from image to orientation parameters, we propose an end-to-end framework that incorporates explicit geometric reasoning. In particular, we design a network that predicts two representations of scene geometry, in both the local camera and global reference coordinate systems, and solves for the camer… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
24
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 35 publications
(24 citation statements)
references
References 48 publications
0
24
0
Order By: Relevance
“…As such, the network fails to produce meaningful estimates for cases where the images are rotated. However, it will be easy to avoid such problems in many practical applications: it is often possible to observe the gravity direction through other sensors or to pre-rotate the image based on geometric cues [78].…”
Section: Discussionmentioning
confidence: 99%
“…As such, the network fails to produce meaningful estimates for cases where the images are rotated. However, it will be easy to avoid such problems in many practical applications: it is often possible to observe the gravity direction through other sensors or to pre-rotate the image based on geometric cues [78].…”
Section: Discussionmentioning
confidence: 99%
“…Calibration methods for only extrinsic parameters have been proposed that are aimed at narrow view cameras [19,32,38,39,44,45] and panoramic 360 • images [10]. These methods cannot calibrate intrinsic parameters, that is, they cannot remove distortion.…”
Section: Related Workmentioning
confidence: 99%
“…Visual cues such as vanishing points in indoor scene images [20,24] can be leveraged to estimate the gravity from images without external sensors such as an IMU. In addition, learning-based methods have employed visual semantics to predict gravity [9,26,35]. This gravity estimate is, in turn, beneficial to recognize visual semantics, e.g., single view depth prediction [8,29].…”
Section: Related Workmentioning
confidence: 99%
“…Notably, Zeng et al [37] and Liao et al [22] proposed to use a L 1 measure on unit vectors and spherical regression loss, respectively, to overcome the limitations of L 2 loss. More recently, UprightNet [35] and VPLNet [32] employed an angular loss (AL), and showed its high effectiveness in both gravity and surface normal predictions, respectively. In this work, we propose a new angular loss called the truncated angular loss that increases robustness to outliers in the training data.…”
Section: Related Workmentioning
confidence: 99%