2023
DOI: 10.1109/tgrs.2023.3250448
|View full text |Cite
|
Sign up to set email alerts
|

Few-Shot Object Detection in Aerial Imagery Guided by Text-Modal Knowledge

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
14
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 20 publications
(19 citation statements)
references
References 64 publications
0
14
0
Order By: Relevance
“…More recently, researchers have explored integrating text data into the visual learning pipeline to improve FSOD performance. Models such as TEMO [38] and TSF-RGR [39] leverage text-modal knowledge extractors to provide prior knowledge on the relationship between base and novel classes, resulting in improved FSOD performance.…”
Section: Few-shot Object Detection In Remote Sensingmentioning
confidence: 99%
“…More recently, researchers have explored integrating text data into the visual learning pipeline to improve FSOD performance. Models such as TEMO [38] and TSF-RGR [39] leverage text-modal knowledge extractors to provide prior knowledge on the relationship between base and novel classes, resulting in improved FSOD performance.…”
Section: Few-shot Object Detection In Remote Sensingmentioning
confidence: 99%
“…To address this limitation, people have been exploring alternative techniques that require fewer RSI annotations. These techniques include semisupervised learning [9,10,19], weakly-supervised learning [20,21,22,23,24], few-shot learning [25,26,27,28,29,30] and active learning [31,32,33],. These methods aim to minimize annotation expenses and speed up the labeling process, making them highly valuable for practical applications in remote sensing.…”
Section: Related Work a Label Efficient Object Detection In Rsismentioning
confidence: 99%
“…Few-shot feature reweighting (FSRW) [5], Meta region-convolutional neural network (R-CNN) [7] aggregate image or region of interest features with support features generated by meta-learners. On the one hand, feature aggregation [9,10] was used, with Xiao et al [9] using class representative features to guide network prediction and Lu et al [10] directing the model with text modality knowledge. On the other hand, feature space enhancement [11,12] was explored, with Zhang et al [11] proposing a metricbased discriminative loss based on transfer learning algorithms and Li et al [12] suggesting a class margin equilibrium method to optimise feature space division and new class reconstruction systematically.…”
Section: Introductionmentioning
confidence: 99%
“…Few‐shot feature reweighting (FSRW) [5], Meta region‐convolutional neural network (R‐CNN) [7] aggregate image or region of interest features with support features generated by meta‐learners. On the one hand, feature aggregation [9, 10] was used, with Xiao et al. [9] using class representative features to guide network prediction and Lu et al.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation