2020
DOI: 10.1007/978-3-030-58598-3_14
|View full text |Cite
|
Sign up to set email alerts
|

RGB-D Salient Object Detection with Cross-Modality Modulation and Selection

Abstract: We present an effective method to progressively integrate and refine the cross-modality complementarities for RGB-D salient object detection (SOD). The proposed network mainly solves two challenging issues: 1) how to effectively integrate the complementary information from RGB image and its corresponding depth map, and 2) how to adaptively select more saliency-related features. First, we propose a cross-modality feature modulation (cmFM) module to enhance feature representations by taking the depth features as… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
84
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 120 publications
(84 citation statements)
references
References 47 publications
0
84
0
Order By: Relevance
“…The LFSD dataset [27] includes 100 RGB-D images via a light field camera. Following [26,33], we adopt 2985 images as our training data, including 1485 samples from NJUD, 700 samples from NLPR, and 800 samples from DUT, and all the remaining images are used as testing. Evaluation metrics.…”
Section: Datasets and Evaluation Metricsmentioning
confidence: 99%
“…The LFSD dataset [27] includes 100 RGB-D images via a light field camera. Following [26,33], we adopt 2985 images as our training data, including 1485 samples from NJUD, 700 samples from NLPR, and 800 samples from DUT, and all the remaining images are used as testing. Evaluation metrics.…”
Section: Datasets and Evaluation Metricsmentioning
confidence: 99%
“…The utilization of RGB-D data for SOD has been extensively explored for years. Traditional methods rely on hand-crafted features [25][26][27][28], while recently, deep learning-based methods have made great progress [5][6][7][8][10][11][12][13][14][15][16][17][18][19][20][21]. Based on the scope of this paper, we divide existing deep-based models into two types according to how they extract RGB and depth features, namely: parallel independent encoders (Fig.…”
Section: Related Workmentioning
confidence: 99%
“…Zhang et al [12] introduced a complimentary interaction module to select useful features. In [13], Li et al enhanced feature representations by taking depth features as priors. Chen et al [14] proposed to extract depth features with a light-weight depth branch and conduct progressive refinement.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations