2021
DOI: 10.1007/s10489-021-02687-7
|View full text |Cite
|
Sign up to set email alerts
|

MMNet: Multi-modal multi-stage network for RGB-T image semantic segmentation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
9

Relationship

0
9

Authors

Journals

citations
Cited by 30 publications
(9 citation statements)
references
References 35 publications
0
9
0
Order By: Relevance
“…e multilanguage machine translation system based on semantic language is composed of two parts, which is a multilanguage unified semantic unit base of high quality, complete, extensible, no discard, no repetition, no false ambiguity, and no normal ambiguity [10].…”
Section: Machine Translation Methods Based On Semantic Unitsmentioning
confidence: 99%
“…e multilanguage machine translation system based on semantic language is composed of two parts, which is a multilanguage unified semantic unit base of high quality, complete, extensible, no discard, no repetition, no false ambiguity, and no normal ambiguity [10].…”
Section: Machine Translation Methods Based On Semantic Unitsmentioning
confidence: 99%
“…1) Comparison on the MFNet Dataset. On the MFNet dataset, we compare our LASNet with 14 state-of-the-art methods, including two RGB semantic segmentation methods (i.e., DANet [12] and HRNet [63]) and their modified RGB-T versions, four RGB-D semantic segmentation methods (i.e., FuseNet [55], D-CNN [59], ACNet [57], and SA-Gate [58]), and eight RGB-T semantic segmentation methods (i.e., MFNet [18], two versions of RTFNet [25], PSTNet [19], MLFNet [26], FuseSeg [28], ABMDRNet [30], MMNet [64], and EGFNet [34]).…”
Section: B Comparison With State-of-the-artsmentioning
confidence: 99%
“…Additionally, the means of these networks that are designed for achieving better performance can also be utilized in single-modality tasks. Modality-specific networks that can aggregate complementary information from different modalities are valuable for multi-modal networks ( Zhu et al, 2016 ; Lan et al, 2022 ). MMFNet uses three specific encoders to separately extract modality-specific features from corresponding modality images.…”
Section: Related Studiesmentioning
confidence: 99%
“…Although some of the abovementioned works utilized modality-specific features to extract additional representative features, the discarded low-level multi-modal fusion features are also crucial in aggregating the complementary information of modalities ( Lan et al, 2022 ). To solve this problem, a multi-modal fusion network is deployed in MSMFF encoder (or decoder) blocks to fuse the modality-specific features and multi-modal fusion features of the former layers.…”
Section: Related Studiesmentioning
confidence: 99%