Proceedings of the 18th ACM International Conference on Multimodal Interaction 2016
DOI: 10.1145/2993148.2993176
|View full text |Cite
|
Sign up to set email alerts
|

Deep multimodal fusion for persuasiveness prediction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
89
0

Year Published

2017
2017
2020
2020

Publication Types

Select...
5
4
1

Relationship

3
7

Authors

Journals

citations
Cited by 165 publications
(89 citation statements)
references
References 19 publications
0
89
0
Order By: Relevance
“…This baseline is trained on the concatenated multimodal features for classification or regression task (Pérez-Rosas et al, 2013), (Park et al, 2014a), (Zadeh et al, 2016b). Deep Fusion The Deep Fusion model (DF) (Nojavanasghari et al, 2016) trains one deep neural model for each modality and then combine the output of each modality network with a joint neural network. Tensor Fusion Network The Tensor Fusion Network (TFN) explicitly models view-specific and cross-view dynamics by creating a multi-dimensional tensor that captures uni-modal, bimodal and trimodal interactions across three modalities.…”
Section: Baseline Modelsmentioning
confidence: 99%
“…This baseline is trained on the concatenated multimodal features for classification or regression task (Pérez-Rosas et al, 2013), (Park et al, 2014a), (Zadeh et al, 2016b). Deep Fusion The Deep Fusion model (DF) (Nojavanasghari et al, 2016) trains one deep neural model for each modality and then combine the output of each modality network with a joint neural network. Tensor Fusion Network The Tensor Fusion Network (TFN) explicitly models view-specific and cross-view dynamics by creating a multi-dimensional tensor that captures uni-modal, bimodal and trimodal interactions across three modalities.…”
Section: Baseline Modelsmentioning
confidence: 99%
“…We used 40% of the data for training, 30% for validation and 30% for testing. We used majority voting, naive Bayes, radial basis function kernel SVM [36], binary logistic regression [37] and a deep neural network [38] as our baseline models. We used the validation set for selecting the hyper parameters of our models.…”
Section: Experimental Methodologymentioning
confidence: 99%
“…Vectorized features from different sources of knowledge can be combined in deep learning using a simple process, such as concatenation or weighted sum, which often has just a few or even number of parameters involved because, the joint training of the deep models will change the layers for high-level extractions of features to compensate for the process needed. Concatenation may be used to combine either low input [6], [7] characteristics or high feature derived from the pre-trained models [8], [9]. Proposed model uses the first technique that is simple operation-based fusion where the vectorized features from images are integrated using concatenation.…”
Section: Related Workmentioning
confidence: 99%