2022
DOI: 10.48550/arxiv.2205.01568
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

RAFT-MSF: Self-Supervised Monocular Scene Flow using Recurrent Optimizer

Abstract: Learning scene flow from a monocular camera still remains a challenging task due to its ill-posedness as well as lack of annotated data. Self-supervised methods demonstrate learning scene flow estimation from unlabeled data, yet their accuracy lags behind (semi-)supervised methods. In this paper, we introduce a self-supervised monocular scene flow method that substantially improves the accuracy over the previous approaches. Based on RAFT, a state-of-the-art optical flow model, we design a new decoder to iterat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(10 citation statements)
references
References 2 publications
0
10
0
Order By: Relevance
“…Hur et al [20] first present a novel selfsupervised model capable of inferring depth and 3D motion field from monocular sequences and surpass the performance of previous multi-task methods. Subsequent studies extend their method into a multi-frame model [21], or employ a recurrent network architecture [2] for better accuracy. Rigidity in Scene Flow.…”
Section: Related Workmentioning
confidence: 99%
See 4 more Smart Citations
“…Hur et al [20] first present a novel selfsupervised model capable of inferring depth and 3D motion field from monocular sequences and surpass the performance of previous multi-task methods. Subsequent studies extend their method into a multi-frame model [21], or employ a recurrent network architecture [2] for better accuracy. Rigidity in Scene Flow.…”
Section: Related Workmentioning
confidence: 99%
“…Spatial Photometric Loss. To address scale ambiguity in monocular scene flow learning, we utilize stereo samples during training as proposed in previous works [20,21,2]. We use the stereoscopic image synthesis loss utilized in [60] to regularize depth estimation on an absolute scale and denote it as L d in our method.…”
Section: Hwmentioning
confidence: 99%
See 3 more Smart Citations