2019
DOI: 10.3390/rs11050594
|View full text |Cite
|
Sign up to set email alerts
|

A Single Shot Framework with Multi-Scale Feature Fusion for Geospatial Object Detection

Abstract: With the rapid advances in remote-sensing technologies and the larger number of satellite images, fast and effective object detection plays an important role in understanding and analyzing image information, which could be further applied to civilian and military fields. Recently object detection methods with region-based convolutional neural network have shown excellent performance. However, these two-stage methods contain region proposal generation and object detection procedures, resulting in low computatio… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
32
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 35 publications
(38 citation statements)
references
References 44 publications
(61 reference statements)
0
32
0
Order By: Relevance
“…The atrous filter enlarges the FOV without increasing the number of parameters to be calculated, and thus saves computational resources. Additionally, high-level features containing semantic information and low level features containing fine details are fused together through up-sampling and concatenation [28] so that different layers of feature are both considered for detection tasks, especially for small object detection. In a similar model [29], the deep residual network (ResNet) is used as the encoder, and high level features are combined with corresponding low-level features as the up-sampling stage decoder.…”
Section: Multi-scalementioning
confidence: 99%
See 1 more Smart Citation
“…The atrous filter enlarges the FOV without increasing the number of parameters to be calculated, and thus saves computational resources. Additionally, high-level features containing semantic information and low level features containing fine details are fused together through up-sampling and concatenation [28] so that different layers of feature are both considered for detection tasks, especially for small object detection. In a similar model [29], the deep residual network (ResNet) is used as the encoder, and high level features are combined with corresponding low-level features as the up-sampling stage decoder.…”
Section: Multi-scalementioning
confidence: 99%
“…The multi-scale concept in this study refers to the relationship between the local and the global, that is, a small local area and a large area within a certain neighborhood. However, the current multi-scale research focuses on the multi-scale feature extraction and fusion of the same training sample image [24][25][26][27][28][29][30][31]. There are few studies on how to consider the semantic analysis of variable regions in different spatial extents [32][33][34][35].…”
Section: Introductionmentioning
confidence: 99%
“…On this basis, object detection in remote sensing imagery has been widely studied in recent years [27][28][29][30][31][32]. In the field of remote sensing, many researchers have made great efforts to object detection methods based on CNN [33][34][35][36][37][38][39].…”
Section: Introductionmentioning
confidence: 99%
“…In addition, the final detection results are obtained by the way of making a decision fusion on the results of the three sub-networks. Based on the YOLOv2 [23], a single-shot geospatial object detection framework based on multi-scale feature fusion modules has been proposed in [33]. Note that detectors in this model are used in conjunction with multi-scale feature fusion modules.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation