2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022
DOI: 10.1109/cvpr52688.2022.01918
|View full text |Cite
|
Sign up to set email alerts
|

Multimodal Material Segmentation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
15
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 21 publications
(35 citation statements)
references
References 20 publications
0
15
0
Order By: Relevance
“…Then, these two results, and the intensity RGB image are fed into three independent conformer encoders 67 and fused using local and global guidance. In a more complex scenario, Liang et al 36 built a network to fuse RGB, infra-red, and polarization cues to produce an outdoor scene segmentation based on the object material type. The proposed pipeline is composed of two core elements: a network that will classify the objects present in one class of a subset of the segmentation labels from the CityScapes dataset 68 and a region-based filter selection module that chooses the modality that provides the most relevant information for determining the type of material of the constitutive elements of each detected object.…”
Section: Image Segmentationmentioning
confidence: 99%
See 3 more Smart Citations
“…Then, these two results, and the intensity RGB image are fed into three independent conformer encoders 67 and fused using local and global guidance. In a more complex scenario, Liang et al 36 built a network to fuse RGB, infra-red, and polarization cues to produce an outdoor scene segmentation based on the object material type. The proposed pipeline is composed of two core elements: a network that will classify the objects present in one class of a subset of the segmentation labels from the CityScapes dataset 68 and a region-based filter selection module that chooses the modality that provides the most relevant information for determining the type of material of the constitutive elements of each detected object.…”
Section: Image Segmentationmentioning
confidence: 99%
“…Mei et al 37 introduced a medium-scale dataset, with 4511 images annotated only for the labels glass and no-glass. Similarly, Liang et al 36 made publicly available a dataset of semantic segmentation of urban scenes, with multi-modal sensors, but it only includes 500 labeled images, and Li et al 39 did it for road segmentation, and with their personalized infra-red polarization camera. Thus, there is a need for a common large-scale benchmark to evaluate the performance of these different segmentation algorithms to trace the direction toward a generalization of the polarization modality.…”
Section: Image Segmentationmentioning
confidence: 99%
See 2 more Smart Citations
“…Additional visual modalities include near-infrared images (Salamati et al, 2014;Liang et al, 2022), thermal images (Ha et al, 2017;Sun et al, 2019d), depth (Wang et al, 2015;Qi et al, 2017;Schneider et al, 2017), surfacenormals (Eigen and Fergus, 2015), 3D LiDAR point clouds (Kim et al, 2018;Jaritz et al, 2018;Caltagirone et al, 2019), etc. .…”
Section: Sis With Additional Modalitiesmentioning
confidence: 99%