2021
DOI: 10.48550/arxiv.2104.03965
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Learning optical flow from still images

Abstract: This paper deals with the scarcity of data for training optical flow networks, highlighting the limitations of existing sources such as labeled synthetic datasets or unlabeled real videos. Specifically, we introduce a framework to generate accurate ground-truth optical flow annotations quickly and in large amounts from any readily available single real picture. Given an image, we use an off-the-shelf monocular depth estimation network to build a plausible point cloud for the observed scene. Then, we virtually … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
0
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(8 citation statements)
references
References 59 publications
(143 reference statements)
0
0
0
Order By: Relevance
“…This is because we add the attention-mechanism module on the decoding side, which significantly reduces the loss of feature information, and the design of the composite loss function addresses the problems of depth loss and the fuzzy target edge, reducing the values of the four error indicators. From the perspective of the threshold accuracy δ, the image prediction accuracy within the three recognized threshold ranges of 1.25, 1.25 2 , and 1.25 3 reached 89.1%, 96.4%, and 98.5%, respectively, which were 0.7%, 0.2% and 0.2% higher than those of the Midas algorithm [41]. In summary, the self-supervised learning method SAU-net outperformed the conventional algorithms for almost all the evaluation indicators.…”
Section: Figure 8 Comparison Of Experimental Resultsmentioning
confidence: 91%
See 4 more Smart Citations
“…This is because we add the attention-mechanism module on the decoding side, which significantly reduces the loss of feature information, and the design of the composite loss function addresses the problems of depth loss and the fuzzy target edge, reducing the values of the four error indicators. From the perspective of the threshold accuracy δ, the image prediction accuracy within the three recognized threshold ranges of 1.25, 1.25 2 , and 1.25 3 reached 89.1%, 96.4%, and 98.5%, respectively, which were 0.7%, 0.2% and 0.2% higher than those of the Midas algorithm [41]. In summary, the self-supervised learning method SAU-net outperformed the conventional algorithms for almost all the evaluation indicators.…”
Section: Figure 8 Comparison Of Experimental Resultsmentioning
confidence: 91%
“…Citation information: DOI 10.1109/ACCESS.2023.3339152 Fig. 8 shows a comparison of the results of the proposed method, Monodepth2 [33], and Midas [41]. Monodepth2 [33]and Midas [41] are classical algorithms for monocular depth estimation.…”
Section: Figure 8 Comparison Of Experimental Resultsmentioning
confidence: 99%
See 3 more Smart Citations