2020
DOI: 10.1007/978-3-030-58558-7_14
|View full text |Cite
|
Sign up to set email alerts
|

Deep Feedback Inverse Problem Solver

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 12 publications
(9 citation statements)
references
References 54 publications
0
9
0
Order By: Relevance
“…We note that differences in the joint labeling schemes used by these monocular 3D methods and our evaluation set do not affect the quality of camera initialization we acquire via rigid alignment, as long as monocular 3D estimates for all views follow the same labeling scheme. Similar to prior work [26], each "neural optimizer step" is trained separately, and stop gradient is applied to all inputs. We used the same architecture across all experiments: L fully-connected 512-dimensional layers followed by a fully-connected 128-dimensional, all with selu nonlinearities [21], followed by a dense output of the size corresponding to the optimization space (flattened 3D pose and weak camera model parameters).…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…We note that differences in the joint labeling schemes used by these monocular 3D methods and our evaluation set do not affect the quality of camera initialization we acquire via rigid alignment, as long as monocular 3D estimates for all views follow the same labeling scheme. Similar to prior work [26], each "neural optimizer step" is trained separately, and stop gradient is applied to all inputs. We used the same architecture across all experiments: L fully-connected 512-dimensional layers followed by a fully-connected 128-dimensional, all with selu nonlinearities [21], followed by a dense output of the size corresponding to the optimization space (flattened 3D pose and weak camera model parameters).…”
Section: Methodsmentioning
confidence: 99%
“…In this final stage, we train an optimizer as a neural network f θ to predict the optimal update dy i+1 to the current guess y i for the pose and cameras, similar to Ma et al [26] for solving inverse problems. Specifically, the update dy i+1 is computed from heatmap mixture parameters g, the current guess y i , the projections of the current guess onto each camera {π (c) (y i )} C c=0 , and the current value of the refinement loss (we omit the dependency of y i on θ in the first line for readability):…”
Section: Neural Optimizer -Stagementioning
confidence: 99%
“…Learning camera pose optimization can be tackled by unrolling the optimizer for a fixed number of steps [21,52,54,83,91,92], computing implicit derivatives [13,15,18,34,68], or crafting losses to mimic optimization steps [88,89]. Multiple works have proposed to learn components of these optimizers [21,52,83], with added complexity and unclear generalization.…”
Section: Related Workmentioning
confidence: 99%
“…Fitting the optimizer to the data: Levenberg-Marquardt is a generic optimization algorithm that involves several heuristics, such as the choice of robust cost function ρ or of the damping factor λ. Past works on learned optimization employ deep networks to predict ρ [52], λ [52,83], or even the pose update δ [21,54], from the residuals and visual features. We argue that this can greatly impair the ability to generalize to new data distributions, as it ties the optimizer to the visual-semantic content of the training data.…”
Section: Direct Alignmentmentioning
confidence: 99%
“…6 DoF pose estimation has a wide range of applications, including augmented reality and robot manipulation [21,22]. Recent progress in differentiable rendering has sparked interest in solving pose estimation via analysis-by-synthesis [4,19,31,48]. However, techniques built around differentiable rendering engines typically require a highquality watertight 3D model of the object for use in rendering.…”
Section: Introductionmentioning
confidence: 99%