2016 Future Technologies Conference (FTC) 2016
DOI: 10.1109/ftc.2016.7821710
|View full text |Cite
|
Sign up to set email alerts
|

Multimodal architecture for emotion in robots using deep learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
8
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
5
3
1

Relationship

1
8

Authors

Journals

citations
Cited by 17 publications
(8 citation statements)
references
References 21 publications
0
8
0
Order By: Relevance
“…Conversely, the use of various sources of information can make the process more resilient against irregularities from the environment (e.g., a noisy room or excessive source of light in the field of view). Albert Mehrabian stated that 7% of the communication information is transferred by language, 38% by paralanguage, and 55% by facial expressions [48] . Through different informative channels, a social robot could be able to receive additional information needed to make its emotion recognition process more sensitive to hidden states of mind.…”
Section: Sensing Modalitiesmentioning
confidence: 99%
“…Conversely, the use of various sources of information can make the process more resilient against irregularities from the environment (e.g., a noisy room or excessive source of light in the field of view). Albert Mehrabian stated that 7% of the communication information is transferred by language, 38% by paralanguage, and 55% by facial expressions [48] . Through different informative channels, a social robot could be able to receive additional information needed to make its emotion recognition process more sensitive to hidden states of mind.…”
Section: Sensing Modalitiesmentioning
confidence: 99%
“…Therefore, many studies have effectively used CNN in facial expression recognition to extract features. To enhance the capabilities of HRI and robot-robot interaction, one study [38] proposed a CNN architecture that gives robots the ability to recognize emotions. The network has three convolution layers and one fully connected layer as the output, and information from speech, gestures, and facial recognition is employed as the CNN input.…”
Section: B Deep Neural Networkmentioning
confidence: 99%
“…Deep learning allows computational models to be composed of multiple processing layers for learning the representations of data with multiple levels of abstraction [11]. Deep learning has been used in object detection [12]- [15], emotion recognition [16], [17], semantic segmentation [18], medical diagnosis [19], and many other domains [20]- [23]. The development of deep learning methods has been steadily widening to include multimodal domains [2].…”
Section: Introductionmentioning
confidence: 99%