2020
DOI: 10.1155/2020/8478527
|View full text |Cite
|
Sign up to set email alerts
|

Emotional Video to Audio Transformation Using Deep Recurrent Neural Networks and a Neuro-Fuzzy System

Abstract: Generating music with emotion similar to that of an input video is a very relevant issue nowadays. Video content creators and automatic movie directors bene t from maintaining their viewers engaged, which can be facilitated by producing novel material eliciting stronger emotions in them. Moreover, there is currently a demand for more empathetic computers to aid humans in applications such as augmenting the perception ability of visually-and/or hearing-impaired people. Current approaches overlook the video's em… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(2 citation statements)
references
References 31 publications
0
2
0
Order By: Relevance
“…e proposed hybrid DNN introduces an adaptive neuro-fuzzy inference method to anticipate a video's mood. Audio waves with a similar emotional experience may be generated using visual signals and a deep short-term memory recurring neural network [8]. Its fuzzy features allow to accurately depicting emotions, while its ability to model data with dynamic time qualities makes it a good fit for data with dynamic temporal properties.…”
Section: Introductionmentioning
confidence: 99%
“…e proposed hybrid DNN introduces an adaptive neuro-fuzzy inference method to anticipate a video's mood. Audio waves with a similar emotional experience may be generated using visual signals and a deep short-term memory recurring neural network [8]. Its fuzzy features allow to accurately depicting emotions, while its ability to model data with dynamic time qualities makes it a good fit for data with dynamic temporal properties.…”
Section: Introductionmentioning
confidence: 99%
“…(vii) DNFS application on video classification and robotics The study of Nguyen et al ( 2019 ) presented a novel convolutional neuro-fuzzy network, which incorporated a CNN into the fuzzy logic domain to derive high-level features of emotions from the text, audio, and image data. Alternatively, the authors in (Cunha Sergio and Lee 2020 ) proposed a novel hybrid DNN with ANFIS to interpret the emotions of a video from its visual features and a deep long short-term memory recurrent neural network to produce the related audio signals with an equal emotional impression. Likewise, in another study (Savchenko et al 2018 ), the authors implemented a fuzzy analysis with a CNN in still-to-video recognition.…”
Section: Analysis and Synthesis Of Datamentioning
confidence: 99%