2018 25th IEEE International Conference on Image Processing (ICIP) 2018
DOI: 10.1109/icip.2018.8451312
|View full text |Cite
|
Sign up to set email alerts
|

On Regression Losses for Deep Depth Estimation

Abstract: Depth estimation from a single monocular image has reached great performances thanks to recent works based on deep networks. However, as various choices of losses, architectures and experimental conditions are proposed in the literature, it is difficult to establish their respective influence on the performances. In this paper we propose an in-depth study of various losses and experimental conditions for depth regression, on NYUv2 dataset. From this study we propose a new network for depth estimation combining… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
43
1
2

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 54 publications
(46 citation statements)
references
References 17 publications
0
43
1
2
Order By: Relevance
“…The NYU-Depth V2 (NYUv2) dataset [35] has approximately 230k pairs of images from 249 scenes for training and 215 scenes for testing. In [16], D3-Net reaches its best performances when trained with the complete dataset. However, NYUv2 also contains a smaller split with 1449 pairs of aligned RGB and depth images, of which 795 pairs are used for training and 654 pairs for testing.…”
Section: Synthetic Nyuv2 With Defocus Blurmentioning
confidence: 99%
See 3 more Smart Citations
“…The NYU-Depth V2 (NYUv2) dataset [35] has approximately 230k pairs of images from 249 scenes for training and 215 scenes for testing. In [16], D3-Net reaches its best performances when trained with the complete dataset. However, NYUv2 also contains a smaller split with 1449 pairs of aligned RGB and depth images, of which 795 pairs are used for training and 654 pairs for testing.…”
Section: Synthetic Nyuv2 With Defocus Blurmentioning
confidence: 99%
“…To perform such tests, we adopt the D3-Net architecture from [16], illustrated in figure 3. We use the PyTorch framework on a NVIDIA TITAN X GPU with 12GB of memory.…”
Section: D3-net Architecturementioning
confidence: 99%
See 2 more Smart Citations
“…A seguir estão apresentadas as funções de custo que foram utilizadas para aprendizado supervisionado, no caso a supervisão é feita por mapas de profundidade. Mais especificamente, empregou-se as funções de custo mse, eigen e berhu para ajustar os parâmetros internos da rede neural profunda apresentada (CARVALHO et al, 2018). A motivação por trás disso foi simplesmente determinar qual seria a função mais adequada.…”
Section: Funções De Custounclassified