2020
DOI: 10.1504/ijiei.2020.105430
|View full text |Cite
|
Sign up to set email alerts
|

Attention-based word-level contextual feature extraction and cross-modality fusion for sentiment analysis and emotion classification

Abstract: Multimodal affective computing has become a popular research area, due to the availability of a large amount of multimodal content. Feature alignment between the modalities and multimodal fusion are the most important issues in multimodal affective computing. To address these issues, the proposed model extracts the features at word-level and forced alignment is used to understand the time-dependent interaction among the modalities. The contextual information among the words of an utterance and between the near… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(4 citation statements)
references
References 25 publications
0
4
0
Order By: Relevance
“…DRCNN may visualize a variety of genome sequence features ( Fig. 5 ) in a compositional manner by using many convolutional layers and filters [ 11 ]. We simulate a set of significant sequences with several motif occurrences in the middle and a set of non-significant sequences with numerous motif occurrences dispersed throughout the sequence.…”
Section: Resultsmentioning
confidence: 99%
“…DRCNN may visualize a variety of genome sequence features ( Fig. 5 ) in a compositional manner by using many convolutional layers and filters [ 11 ]. We simulate a set of significant sequences with several motif occurrences in the middle and a set of non-significant sequences with numerous motif occurrences dispersed throughout the sequence.…”
Section: Resultsmentioning
confidence: 99%
“…A possible solution could be cross-modality attention, where we may carry out feature extraction at common frequency domains as suggested in Huddar et al . ( 2020 ). After that, we can align the features of the two modalities in the time dimension and feed them to the cross-modality models (such as Wei et al ., 2020 ).…”
Section: Discussionmentioning
confidence: 99%
“…In this work, we execute each tree that terminates down in the bottom of the leaf with a new subset of predictors and we acquire a new observation subset vi. We proceed to the next tree, which has a slightly different leaf, which is anticipated to be of a different set of significant classes [5].…”
Section: Figure42 Proposed Rf Modelmentioning
confidence: 99%