2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2020
DOI: 10.1109/iros45743.2020.9340821
|View full text |Cite
|
Sign up to set email alerts
|

Semantic Segmentation of Underwater Imagery: Dataset and Benchmark

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
81
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 108 publications
(81 citation statements)
references
References 26 publications
0
81
0
Order By: Relevance
“…These tasks are mainly designed to extract knowledge from underwater videos and images. Despite the recent use of CNNs for various visual analysis tasks such as segmentation [55][56][57][58], localisation [59][60][61], and counting [62][63][64], the most common and the widest studied CV task in underwater fish habitat monitoring has been classification. Therefore, in this paper, we focus mainly on classification of underwater fish images.…”
Section: Applications Of Deep Learning In Fish-habitat Monitoringmentioning
confidence: 99%
“…These tasks are mainly designed to extract knowledge from underwater videos and images. Despite the recent use of CNNs for various visual analysis tasks such as segmentation [55][56][57][58], localisation [59][60][61], and counting [62][63][64], the most common and the widest studied CV task in underwater fish habitat monitoring has been classification. Therefore, in this paper, we focus mainly on classification of underwater fish images.…”
Section: Applications Of Deep Learning In Fish-habitat Monitoringmentioning
confidence: 99%
“…Obviously, this simple network design is outdated and ineffective, and requires very large computational resources. The work [21] proposes an optional residual skip block consisting of three convolutional layers with batch normalisation and ReLU non‐linearity after each convolutional layer. Furthermore, they embed this residual skip block module in an end‐to‐end architecture to extract multi‐scale features.…”
Section: Related Workmentioning
confidence: 99%
“…As its low-resolution predictions generally result in poor performance, we create a higher-resolution version by incorporating two ideas from [10,70]. First, we [20] Consumer 22 10K ADE20K [76] Consumer 150 20K COCO Panoptic [9,34,43] Consumer 134 118K KITTI [1] Driving 30 150 CamVid [7] Driving 23 367 CityScapes [14] Driving 33 3K India Driving Dataset (IDD) [64] Driving 35 7K Berkeley Deep Drive (BDD) [74] Driving 20 7K Mapillary Vista Dataset [51] Driving 66 18K ISPRS [57] Aerial 6 4K iSAID [68,71] Aerial 16 27K SUN RGB-D [60] Indoor 37 5K ScanNet [15] Indoor 41 19K SUIM [33] Underwater 8 1525 vKITTI2 [8,23] Synthetic driving 9 43K vGallery [69] Synthetic indoor 8 44K…”
Section: Model Architecturesmentioning
confidence: 99%