2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) 2018
DOI: 10.1109/cvprw.2018.00047
|View full text |Cite
|
Sign up to set email alerts
|

Stacked U-Nets for Ground Material Segmentation in Remote Sensing Imagery

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
33
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 55 publications
(33 citation statements)
references
References 18 publications
0
33
0
Order By: Relevance
“…In recent years, some tentative work has been proposed for multimodal data analysis in RS ( Gómez-Chova et al, 2015 , Kampffmeyer et al, 2016 , Máttyus et al, 2016 , Audebert et al, 2016 , Audebert et al, 2017 , Zampieri et al, 2018 , Ghosh et al, 2018 ). Related to ours for scene parsing with multimodal deep networks, an early deep fusion architecture, simply stacking all multi-modalities as input, is used for semantic segmentation of urban RS images ( Kampffmeyer et al, 2016 ).…”
Section: Related Workmentioning
confidence: 99%
“…In recent years, some tentative work has been proposed for multimodal data analysis in RS ( Gómez-Chova et al, 2015 , Kampffmeyer et al, 2016 , Máttyus et al, 2016 , Audebert et al, 2016 , Audebert et al, 2017 , Zampieri et al, 2018 , Ghosh et al, 2018 ). Related to ours for scene parsing with multimodal deep networks, an early deep fusion architecture, simply stacking all multi-modalities as input, is used for semantic segmentation of urban RS images ( Kampffmeyer et al, 2016 ).…”
Section: Related Workmentioning
confidence: 99%
“…4 showing the end-to-end pipeline of the DDCM-Net combined with a pre-trained model for land cover classification. Compared to other encoder-decoder architectures, our proposed DDCM-Net only fuses low-level features one time before the final prediction CNN layers, instead of aggregating multi-scale features captured at many different encoder layers [2], [6], [27], [31], [32], [33], [34], [35], [36]. This makes our model simple and neat, yet effective with lower computational cost.…”
Section: The Ddcm Networkmentioning
confidence: 99%
“…Our DDCM-SER50 model achieves new state-of-the-art result with 56.2% mIoU on DeepGlobe land cover classification challenge dataset. As shown in Table IV, we compare our DDCM network with other published models ( [30], [31], [32], [33], [34], [35], [36]) on the hold-out validation set (the public leaderboard 3 up to the date of May 1, 2019). Our model obtained above 3.5% higher mIoU than the second best model [35].…”
Section: Deepglobementioning
confidence: 99%
“…Through end-to-end training, the U-Net takes on as input an image of any size and produces a segmentation map of similar dimensions. Thus, due to these enhanced properties, U-Net gained a high level of success and has been applied in various segmentation tasks [4], [5].…”
Section: Introductionmentioning
confidence: 99%