2018
DOI: 10.1016/j.isprsjprs.2017.11.003
|View full text |Cite
|
Sign up to set email alerts
|

MugNet: Deep learning for hyperspectral image classification using limited samples

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
88
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 213 publications
(88 citation statements)
references
References 53 publications
0
88
0
Order By: Relevance
“…A trainable block, called the field-of-view (FoV), is proposed in [28] to boost the performance of the FCN. With the successful applications of U-Net in the pixel-wise area labellings, most current models [28][29][30][31][32][33] use encoder-decoder architectures. The mutation models enhance the buildings' semantic boundaries Remote Sens.…”
mentioning
confidence: 99%
See 1 more Smart Citation
“…A trainable block, called the field-of-view (FoV), is proposed in [28] to boost the performance of the FCN. With the successful applications of U-Net in the pixel-wise area labellings, most current models [28][29][30][31][32][33] use encoder-decoder architectures. The mutation models enhance the buildings' semantic boundaries Remote Sens.…”
mentioning
confidence: 99%
“…Audebert et al [32] proposed an efficient multi-scale approach to leverage both a large spatial context and the high-resolution data and investigated the early and late fusion of Lidar and multispectral data to cover the scale variance of buildings from different areas. In [33,34], the extra geographical information (DSM, DEM, and Lidar images) are fed into a carefully designed FCN, together with high-resolution RGB images, and the results indicate that abundant features always lead to sharper predicted building boundaries. Moreover, post-processing methods, such as Guider Filter [1] and Conditional Random Field (CRF) methods [35,36], have been heavily researched and attempted to preserve the structure consistency between the building predictions and the original images.…”
mentioning
confidence: 99%
“…Some studies lightened the network by utilizing the 1×1 convolutional layers [14], [15]. [16] reduced the parameters by designing a simplified four layers network without hyperparameters, which applied an ensemble manner to utilize spatial-spectral information from hyperspectral data using 20 labeled samples per class. However, a deeper network with more parameters is expected to have a better performance.…”
Section: Introductionmentioning
confidence: 99%
“…Recently, deep learning methods have made significant progress in many computer visual tasks [38][39][40][41], and the cloud detection methods based on deep learning have attracted attention [42,43]. Shi et al [44] and Goff et al [45] used superpixel segmentation and deep Convolutional Neural Networks (CNNs) to detect clouds from Quickbird and Google Earth images and SPOT 6 images, respectively.…”
Section: Introductionmentioning
confidence: 99%