2022
DOI: 10.1109/tcsvt.2021.3072207
|View full text |Cite
|
Sign up to set email alerts
|

SiamCDA: Complementarity- and Distractor-Aware RGB-T Tracking Based on Siamese Network

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
20
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
2
2

Relationship

0
10

Authors

Journals

citations
Cited by 62 publications
(20 citation statements)
references
References 54 publications
0
20
0
Order By: Relevance
“…Zhu et al [ 7 ] proposed a trident architecture to integrate the fused modality features and two modality-specific features, thus achieving robust target representation. Zhang et al [ 33 ] introduced a complementary perception module for multi-modal feature fusion, which reduces the modality discrepancy between single-modal features to enhance the discriminability of fused features. This is done to fully utilize training data and address various challenges such as illumination variations, occlusion, thermal crossover, and fast motion.…”
Section: Related Workmentioning
confidence: 99%
“…Zhu et al [ 7 ] proposed a trident architecture to integrate the fused modality features and two modality-specific features, thus achieving robust target representation. Zhang et al [ 33 ] introduced a complementary perception module for multi-modal feature fusion, which reduces the modality discrepancy between single-modal features to enhance the discriminability of fused features. This is done to fully utilize training data and address various challenges such as illumination variations, occlusion, thermal crossover, and fast motion.…”
Section: Related Workmentioning
confidence: 99%
“…Gao et al [ 33 ] weighted the modals to make the network focus on more favorable fields to effectively integrate different modals. Zhang et al [ 34 ] took the feature maps from two-stream Siamese networks as inputs and weighted the features through the weight generation sub-network to obtain the additional information between modals. Then, the enhanced features were obtained by using cross-modal residual connections, and finally, these features were concatenated.…”
Section: Related Workmentioning
confidence: 99%
“…DeT [47] adds a depth feature extraction branch to the original ATOM [7] or DiMP [3] tracker and fine-tunes on RGB-D training data. Zhang et al [57] extend SiamRPN++ [21] with dual-modal inputs for RGB-T tracking. They first con-struct a unimodal tracking network trained on RGB data, then tune the whole extended multi-modal network with RGB-T image pairs.…”
Section: Introductionmentioning
confidence: 99%