2015 28th SIBGRAPI Conference on Graphics, Patterns and Images 2015
DOI: 10.1109/sibgrapi.2015.39
|View full text |Cite
|
Sign up to set email alerts
|

Improving Spatial Feature Representation from Aerial Scenes by Using Convolutional Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
27
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
5
4

Relationship

2
7

Authors

Journals

citations
Cited by 43 publications
(29 citation statements)
references
References 24 publications
0
27
0
Order By: Relevance
“…PatreoNet, presented in [15], is a network capable of learning specific spatial features from remote sensing images, without any pre-processing step or descriptor evaluation. This network, which architecture can be seen in Figure 7a, has 3 convolutional layers, 3 pooling ones and 3 fully-connected ones (considering the softmax).…”
Section: Patreonetmentioning
confidence: 99%
See 1 more Smart Citation
“…PatreoNet, presented in [15], is a network capable of learning specific spatial features from remote sensing images, without any pre-processing step or descriptor evaluation. This network, which architecture can be seen in Figure 7a, has 3 convolutional layers, 3 pooling ones and 3 fully-connected ones (considering the softmax).…”
Section: Patreonetmentioning
confidence: 99%
“…Recently, deep learning has become the new state-of-the-art solution for visual recognition. Given its success, deep learning has been intensively used in several distinct tasks of different domains [12,13], including remote sensing [14,15,16]. In remote sensing, the use of deep learning is growing very quickly, since it has a natural ability to effectively encode spectral and spatial information based mainly on the data itself.…”
Section: Introductionmentioning
confidence: 99%
“…It can be also noted that the proposed method outperforms other deep models, such as GoogLeNet [7] which obtains 97.10%, VGG-VD16-1 st -FC+Aug [23] which obtains 96.88%, SPP-net+MKL [29] which obtains 96.38%, and MCNN [21] which obtains 96.66%. For Brazilian Coffee Scene dataset, the proposed method also obtains 91.24% outperforms 88.46% which is obtained by LQPCANet [37], 85.36% by VGG16 [22], 89.79% by ConvNet [38], 90.94% by CaffeNet [7] and 91.13% by D-DSML-CaffeNet [9]. For Google dataset, the proposed method obtains 92.04% which is better than 82.81% by TF-CNN [39], 89.88% by RDSG-CNN [39], 87.68% by Fine-tuned CaffeNet.…”
Section: Comparisons With the Most Recent Methodsmentioning
confidence: 86%
“…Methods from this area, commonly represented as multilayered neural networks, are able to learn both the features and the classifier in a unified manner, adjusting themselves to better represent the characteristics of the data and their labels. A specific deep learning method, called Convolutional (Neural) Networks (ConvNets) [9], is the most popular for learning visual features in computer vision applications, including remote sensing [10], [11]. This type of network relies on the natural stationary property of an image, i.e., the statistics of one part of the image are assumed to be the same as those of any other part.…”
Section: Introductionmentioning
confidence: 99%