Depth modal features can provide complementary information for salient object detection (SOD). Most of the existing RGB-D SOD methods focus on fully combining RGB and Depth modal features without distinguishing them. In this paper, we propose a new depth guided cross-modal residual adaptive network for RGB-D SOD. We use two independent resnet-50 to extract the features of the two modes respectively. Then the cross-modal channel-wise refinement module is designed to obtain complementary modal information. We design a crossmodal guided module to make complementary modal information guide RGB image feature extraction. Finally, the residual adaptive selection module is used to enhance the spatial mutual concerns between the two modal features to achieve multimodal information fusion. Experimental results show that our method can achieve a more reasonable fusion state of RGB and Depth, and verify the effectiveness of our final saliency model.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.