2020 IEEE 23rd International Conference on Information Fusion (FUSION) 2020
DOI: 10.23919/fusion45008.2020.9190246
|View full text |Cite
|
Sign up to set email alerts
|

Early vs Late Fusion in Multimodal Convolutional Neural Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
43
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 90 publications
(52 citation statements)
references
References 27 publications
1
43
0
Order By: Relevance
“…Although various types of CNN structures can be applied to radar signals, it is more suitable to predict anthropometric parameters by applying each radar signal to an independent convolution layer rather than attaching all radar signals to one image and using it as input. The datasets created by each radar sensor have characteristics of human signals that depend on the location where it is installed, and since three sensors are used, a late-fusion method, rather than feature-level fusion, is used for the CNN structure, given the multi modality 22,23 .…”
Section: Height Estimation Using Radar Signal Processingmentioning
confidence: 99%
“…Although various types of CNN structures can be applied to radar signals, it is more suitable to predict anthropometric parameters by applying each radar signal to an independent convolution layer rather than attaching all radar signals to one image and using it as input. The datasets created by each radar sensor have characteristics of human signals that depend on the location where it is installed, and since three sensors are used, a late-fusion method, rather than feature-level fusion, is used for the CNN structure, given the multi modality 22,23 .…”
Section: Height Estimation Using Radar Signal Processingmentioning
confidence: 99%
“…The co-existence of diverse modes of input information naturally raises the question of their combination. That is, a framework facilitating the collaboration of the multiple input representations is needed in order to profit from every type of information and provide an enhanced final prediction [19]. Two are the state-of-the-art techniques that can help towards this direction: a) model ensembling through stacked generalization and b) feature concatenation inside a CNN.…”
Section: Modality Fusion Techniquesmentioning
confidence: 99%
“…Feature concatenation [18,19] on the other hand can be performed inside a CNN: that is, two distinct branches, respectively extract features from the tabular data and the images. These features are then concatenated and fed into the CNN's regressor part, which performs the predictions.…”
Section: Modality Fusion Techniquesmentioning
confidence: 99%
See 1 more Smart Citation
“…We combine the two types of features by concatenating the vectors representing each aspect in a single vector, used as input to a supervised classification algorithm. According to [Gadzicki et al 2020], this approach of combining multi-modal features is referred to as early-maturing fusion.…”
Section: Feature Extraction From News and Diffusion Network For Fake News Classificationmentioning
confidence: 99%