2021 IEEE Winter Conference on Applications of Computer Vision (WACV) 2021
DOI: 10.1109/wacv48630.2021.00389
|View full text |Cite
|
Sign up to set email alerts
|

Self-supervised Visual-LiDAR Odometry with Flip Consistency

Abstract: Most learning-based methods estimate ego-motion by utilizing visual sensors, which suffer from dramatic lighting variations and textureless scenarios. In this paper, we incorporate sparse but accurate depth measurements obtained from lidars to overcome the limitation of visual methods. To this end, we design a self-supervised visual-lidar odometry (Self-VLO) framework. It takes both monocular images and sparse depth maps projected from 3D lidar points as input, and produces pose and depth estimations in an end… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
30
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 23 publications
(30 citation statements)
references
References 46 publications
0
30
0
Order By: Relevance
“…Instead of direct regression model for ego-motion, [26] proposes to estimate an optical flow between two images and solves ego-motion as an optimal fundamental matrix. Besides these works which still perform ego-motion estimation purely in the RGB domain, some other works attempt to introduce data from other sensors, e.g., LiDAR [3] and IMU [14,16], as additional inputs to improve the ego-motion accuracy in the spirit of sensor fusion. Following the idea of performing motion estimation in a mixed domain, [8] further proposes a two-stream network that leverages the original RGB images and the internally inferred depth maps as inputs of egomotion network.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…Instead of direct regression model for ego-motion, [26] proposes to estimate an optical flow between two images and solves ego-motion as an optimal fundamental matrix. Besides these works which still perform ego-motion estimation purely in the RGB domain, some other works attempt to introduce data from other sensors, e.g., LiDAR [3] and IMU [14,16], as additional inputs to improve the ego-motion accuracy in the spirit of sensor fusion. Following the idea of performing motion estimation in a mixed domain, [8] further proposes a two-stream network that leverages the original RGB images and the internally inferred depth maps as inputs of egomotion network.…”
Section: Related Workmentioning
confidence: 99%
“…One key component in them is to get an accurate ego-motion between two consecutive frames, which is often carried as camera pose estimation using RGB images. On top of the great success of classical methods based on 3D geometry and camera models, learning-based pose estimation methods [1,2] have recently got increasing research interests for their good fits to training data and feasibility in typical severe situations, e.g., poor lighting [3]. Most of these works treat the pose estimation problem as a regression from input color images, and design pose estimators based on the convolutional neural network (CNN) [1,2,[4][5][6][7][8] or the recurrent neural network (RNN) [9][10][11][12].…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…Others combine features from different sources or sensors. [12] uses a Siamese pyramid network which combines features from two pairs of sparse depth and RGB images and regresses the relative motion, and depth.…”
Section: B Deep Lidar Odometriesmentioning
confidence: 99%