2019
DOI: 10.3390/rs11141687
|View full text |Cite
|
Sign up to set email alerts
|

Combing Triple-Part Features of Convolutional Neural Networks for Scene Classification in Remote Sensing

Abstract: High spatial resolution remote sensing (HSRRS) images contain complex geometrical structures and spatial patterns, and thus HSRRS scene classification has become a significant challenge in the remote sensing community. In recent years, convolutional neural network (CNN)-based methods have attracted tremendous attention and obtained excellent performance in scene classification. However, traditional CNN-based methods focus on processing original red-green-blue (RGB) image-based features or CNN-based single-laye… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
23
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 56 publications
(25 citation statements)
references
References 79 publications
0
23
0
Order By: Relevance
“…Therefore, we only need to modify the structure of CNN architectures rather than manually design the features, which can save the human resources. For another, deep features are more discriminative in distinguishing similar scene categories compared with the other two types of features [39].…”
Section: R E T R a C T E Dmentioning
confidence: 99%
“…Therefore, we only need to modify the structure of CNN architectures rather than manually design the features, which can save the human resources. For another, deep features are more discriminative in distinguishing similar scene categories compared with the other two types of features [39].…”
Section: R E T R a C T E Dmentioning
confidence: 99%
“…The definitions of are the same as those in Equation (1). Compared with the widely used cross-entropy loss, L1 loss is more sensitive to the variation in the local information, which is critical to learning the structural information.…”
Section: Cyclenet Architecturementioning
confidence: 99%
“…Similar to the refining stage of the classical perceptual loss, the ground truth G mask and the predicted building map S from the building extraction network θ are separately fed into the pretrained CycleNet η. With the input of S (G mask ), the extracted features from the nth convolutional layer in both the boundary block and mask block are represented by ∅ n , ϕ n (∅ n ,φ n ), respectively, and n ∈ [1,4]. The transfer loss functions are used to minimize the error between ∅ n , ϕ n and∅ n ,φ n .…”
Section: Transfer Loss Functionsmentioning
confidence: 99%
See 2 more Smart Citations