2022
DOI: 10.1016/j.displa.2022.102162
|View full text |Cite
|
Sign up to set email alerts
|

Cross attention redistribution with contrastive learning for few shot object detection

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 15 publications
(13 citation statements)
references
References 43 publications
0
13
0
Order By: Relevance
“…(1) LSTD [58] proposes a regularization method based on transfer knowledge and background depression regularizations to enhance the fine-tuning effect; (2) Meta YOLO [20] learns feature representation by reweighting module to reassign feature weights; (3) MetaDet [59] solves the problem of few-shot classification and localization simultaneously through a weight prediction meta-model; (4) CME [21] balances the novel class margins by class margin loss and feature interference; (5) Meta R-CNN [25] obtains the class attention vector through the predictor-head remodeling network (PRN) module to remodel the ROI feature; (6) Viewpoint [26] performs efficient feature similarity calculation through feature subtraction; (7) DCNet [27] introduces adaptive context awareness into the feature aggregation module, to gain better global features and local features; (8) FSCN [60] introduces a novel few-shot classification refinement mechanism to improve the final classification; (9) FsDetView+ISAM+QSAM [61] generated an individual prototype for each sample to extract the unique characteristics of each support sample; (10) CAReD [62] maximizes the inter-class distance and minimizes the intra-class distance through contrastive learning.…”
Section: Baseline Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…(1) LSTD [58] proposes a regularization method based on transfer knowledge and background depression regularizations to enhance the fine-tuning effect; (2) Meta YOLO [20] learns feature representation by reweighting module to reassign feature weights; (3) MetaDet [59] solves the problem of few-shot classification and localization simultaneously through a weight prediction meta-model; (4) CME [21] balances the novel class margins by class margin loss and feature interference; (5) Meta R-CNN [25] obtains the class attention vector through the predictor-head remodeling network (PRN) module to remodel the ROI feature; (6) Viewpoint [26] performs efficient feature similarity calculation through feature subtraction; (7) DCNet [27] introduces adaptive context awareness into the feature aggregation module, to gain better global features and local features; (8) FSCN [60] introduces a novel few-shot classification refinement mechanism to improve the final classification; (9) FsDetView+ISAM+QSAM [61] generated an individual prototype for each sample to extract the unique characteristics of each support sample; (10) CAReD [62] maximizes the inter-class distance and minimizes the intra-class distance through contrastive learning.…”
Section: Baseline Methodsmentioning
confidence: 99%
“…Compared with FsDetView+ISAM+QSAM [61], BFR achieves 5.63% average improvements in mAP. Compared with CAReD [62], BFR achieves 2.11% average improvements in mAP.…”
Section: Voc Datasetmentioning
confidence: 99%
See 1 more Smart Citation
“…Quan et al [52] (CAReD) followed a similar approach. However, the weight w i is determined by the softmax over the correlation between the support features f S,c i and all other support features { f S,c j } K j=1 of the same category c. Due to the softmax, the weighting factors already sum up to 1 and the factor (1/k) is omitted.…”
Section: ) Aggregation Of Several Support Imagesmentioning
confidence: 99%
“…The outputs of all three matching modules are summed to give the final matching score. Many others [52], [66], [68], [69] adopt or build upon this multirelation detector. The additionally proposed two-way contrastive training strategy is implemented as follows.…”
Section: E Increase Discriminative Powermentioning
confidence: 99%