2022
DOI: 10.1109/tcsvt.2021.3077058
|View full text |Cite
|
Sign up to set email alerts
|

ECFFNet: Effective and Consistent Feature Fusion Network for RGB-T Salient Object Detection

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
29
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
3
1

Relationship

0
10

Authors

Journals

citations
Cited by 143 publications
(29 citation statements)
references
References 64 publications
0
29
0
Order By: Relevance
“…However, we simply concatenate the multilevel feature maps in the RF module as a feature fusion procedure, where the relationship among these maps is seldom considered. As we know, various fusion approaches, aiming to flexibly exploit the feature information, have successfully been presented in different scenarios [46–48]. Moreover, an attention mechanism has also been proved as a useful fusion strategy in image denoising as mentioned before.…”
Section: Discussionmentioning
confidence: 99%
“…However, we simply concatenate the multilevel feature maps in the RF module as a feature fusion procedure, where the relationship among these maps is seldom considered. As we know, various fusion approaches, aiming to flexibly exploit the feature information, have successfully been presented in different scenarios [46–48]. Moreover, an attention mechanism has also been proved as a useful fusion strategy in image denoising as mentioned before.…”
Section: Discussionmentioning
confidence: 99%
“…Most SOD methods [37,9,38,39,11,12] mainly focus on solving the integration of multi-level features to generate a salient map with accurate location and internal consistency. Other works are devoted to studying how to use multi-modal information, such as depth cues [40,41,42] or thermal infrared cues [43,44], as auxiliary inputs of the models to alleviate the defects of individual RGB sources, such as low-light environments and similar texture scenes. It is noted that, some methods [13,17,10,33] also introduce the edge detection as a joint or auxiliary task for SOD, which make the models pay more attention to the object structure and thus improve the localization of salient objects.…”
Section: Fully Supervised Salient Object Detectionmentioning
confidence: 99%
“…Whereas, methods such as MFNet [62], RTFNet [63], PST900 [64], FuseSeg [65] combine the potential of RGB images along with thermal images using CNN architectures for semantic segmentation of outdoor scenes, providing accurate segmentation results even in presence of degraded lighting conditions. Authors of ECFFNet [66], perform the fusion of RGB and thermal images at feature level, which provides a complementary information effectively improving the object detection in different lighting conditions. Authors in [67] and [68] perform a fusion of RGB, depth and thermal camera computing descriptors in all the three image spaces and fusing them in a weighted average manner for efficient human detection.…”
Section: A Scene Understandingmentioning
confidence: 99%