2019
DOI: 10.1109/lsens.2018.2880790
|View full text |Cite
|
Sign up to set email alerts
|

Learning Fused Representations for Large-Scale Multimodal Classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 11 publications
(3 citation statements)
references
References 9 publications
0
3
0
Order By: Relevance
“…We trained a custom CNN on the image-like embeddings and performed a grid search to tune its hyperparameters. The approach of encoding and stacking multimodal features into a single source suitable for training CNNs was inspired by Nawaz et al who fused image and text embeddings to improve classification performance [50].…”
Section: Treatment Prediction Using Deep Neural Networkmentioning
confidence: 99%
“…We trained a custom CNN on the image-like embeddings and performed a grid search to tune its hyperparameters. The approach of encoding and stacking multimodal features into a single source suitable for training CNNs was inspired by Nawaz et al who fused image and text embeddings to improve classification performance [50].…”
Section: Treatment Prediction Using Deep Neural Networkmentioning
confidence: 99%
“…Several methods have been proposed to fuse image and textual features. Encoded textual features to the image domain and passing the resulting image through a CNN have been shown to improve classification performance [63][64][65]. Furthermore, transformer-based models have been successfully combined for multimodal classification [66].…”
Section: Multimodal Classificationmentioning
confidence: 99%
“…Clustering means that the samples with large similarity in data are gathered into a category in terms of similarity criterion and express as a local area in feature space [43], [44]. It can discover and mine the arbitrary shape class clusters in remote sensing data and has great potential for classification of remote sensing data [45]- [47].…”
Section: Introductionmentioning
confidence: 99%