2021
DOI: 10.48550/arxiv.2106.04067
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

LocalTrans: A Multiscale Local Transformer Network for Cross-Resolution Homography Estimation

Abstract: Our LocalTrans Coventional Feature-based Method [42] * Equal contribution MS-COCO dataset and the real-captured cross-resolution dataset show that the proposed network outperforms existing state-of-the-art feature-based and deep-learning-based homography estimation methods, and is able to accurately align images under 10× resolution gap.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
4
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(4 citation statements)
references
References 43 publications
(124 reference statements)
0
4
0
Order By: Relevance
“…Deep Homography Estimation Deep homography estimation can be categorized into supervised and unsupervised methods. Supervised methods [8,18,28] learn from image pairs with ground truth homographies, which are difficult to obtain for natural images in the wild. If learning from synthetic images, the lack of realistic transformation will degrade their generalization ability.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…Deep Homography Estimation Deep homography estimation can be categorized into supervised and unsupervised methods. Supervised methods [8,18,28] learn from image pairs with ground truth homographies, which are difficult to obtain for natural images in the wild. If learning from synthetic images, the lack of realistic transformation will degrade their generalization ability.…”
Section: Related Workmentioning
confidence: 99%
“…[39] and [18] introduced mask prediction into homography estimation, but their goal is to remove large foregrounds or moving objects, while our goal is to preserve a single dominant plane with explicit constraint. Recently, Shao et al [28] proposed a supervised transformer for cross-resolution homography estimation. However, aiming at different tasks, our architecture designs are also different, where they propose a transformer with local attention, while ours contains a self-attention encoder and class-attention decoder.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations