2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.01000
|View full text |Cite
|
Sign up to set email alerts
|

Refine and Distill: Exploiting Cycle-Inconsistency and Knowledge Distillation for Unsupervised Monocular Depth Estimation

Abstract: Nowadays, the majority of state of the art monocular depth estimation techniques are based on supervised deep learning models. However, collecting RGB images with associated depth maps is a very time consuming procedure. Therefore, recent works have proposed deep architectures for addressing the monocular depth prediction task as a reconstruction problem, thus avoiding the need of collecting ground-truth depth. Following these works, we propose a novel self-supervised deep model for estimating depth maps. Our … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
62
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
3
2

Relationship

1
9

Authors

Journals

citations
Cited by 134 publications
(62 citation statements)
references
References 40 publications
(112 reference statements)
0
62
0
Order By: Relevance
“…Specifically, we report in the table competitors for whose the source code or trained models are available, self-supervised either using monocular or stereo images. Unfortunately, code and models are not available for [46,34], and hence, we are not able to compare with them. Methods marked with * have been pre-trained on CityScapes dataset [5], for whose the authors do not provide weights trained on KITTI only.…”
Section: Ablation Studiesmentioning
confidence: 99%
“…Specifically, we report in the table competitors for whose the source code or trained models are available, self-supervised either using monocular or stereo images. Unfortunately, code and models are not available for [46,34], and hence, we are not able to compare with them. Methods marked with * have been pre-trained on CityScapes dataset [5], for whose the authors do not provide weights trained on KITTI only.…”
Section: Ablation Studiesmentioning
confidence: 99%
“…The work of [24] integrates heterogeneous predictions from global and local networks. Andrea Pilzer et.al [34] train a student network to predict a disparity map and a backward cycle network for generating image to re-synthesize back the input image. Then the third network exploits the inconsistency between the original and the reconstructed input frame in order to output a refined depth map and knowledge distillation are exploited.…”
Section: B Benchmark Performance 1) Compared Methodsmentioning
confidence: 99%
“…The purpose of knowledge distillation is to transfer the knowledge learned by the teacher network to the student network, so that the functions learned by the large model are compressed into smaller and faster models. Pilzer et al [72] followed this idea, and knowledge distillation is used to transfer information from the refinement network to the student network. Considering the effectiveness of training with synthetic images, Zhao et al [78] adopted the framework of cycle GAN for the transformation between the synthetic and real domains to expand the data set, and propose a geometryaware symmetric domain adaptation network (GASDA) to make better use of the synthetic data.…”
Section: Semi-supervised Monocular Depth Estimationmentioning
confidence: 99%