2016
DOI: 10.14257/ijsh.2016.10.8.14
|View full text |Cite
|
Sign up to set email alerts
|

Smart Home Entertainment System with Personalized Recommendation and Speech Emotion Recognition Support

Abstract: Abstract

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
3
0
1

Year Published

2016
2016
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 9 publications
0
3
0
1
Order By: Relevance
“…Emotion detection using neural networks.-Voice can be characterized by various parameters such as pitch (indicating the level of highness/lowness of a tone) and frequency (indicating the variation in the pitch) which are useful for determining the emotion of a speaker. Building on earlier research on voice recognition (e.g., Pan et al, 2012;Gao et al, 2017;Likitha et al, 2017;Bhavan et al, 2019) MFCCs can be derived from Mel Spectrogram Frequencies, we find that using both types of features helps to improve the accuracy of the model. Note that the number of Mel spectrogram coefficients, MFCCs, and chroma coefficients can be adjusted to achieve more accurate predictions.…”
Section: A Voice Tonementioning
confidence: 59%
“…Emotion detection using neural networks.-Voice can be characterized by various parameters such as pitch (indicating the level of highness/lowness of a tone) and frequency (indicating the variation in the pitch) which are useful for determining the emotion of a speaker. Building on earlier research on voice recognition (e.g., Pan et al, 2012;Gao et al, 2017;Likitha et al, 2017;Bhavan et al, 2019) MFCCs can be derived from Mel Spectrogram Frequencies, we find that using both types of features helps to improve the accuracy of the model. Note that the number of Mel spectrogram coefficients, MFCCs, and chroma coefficients can be adjusted to achieve more accurate predictions.…”
Section: A Voice Tonementioning
confidence: 59%
“…Tổ hợp các hệ số MFCC, LPCC (Linear Predictive Cepstral Coefficients), RASTA PLP (Relative Spectral Transform -Perceptual Linear Prediction) và các hệ số logarit của công suất đối với tần số đã được xem là tập các đặc điểm để phân loại các cảm xúc: tức giận, chán, bình thường, vui, buồn trong tiếng phổ thông Trung Quốc [11]. SVM cũng được dùng để nhận dạng 3 cảm xúc vui, buồn, bình thường của tiếng Trung Quốc [16] sử dụng các tham số như năng lượng, tần số cơ bản, LPCC, MFCC và MEDC (Mel-Energy spectrum Dynamic Coefficients). [17] sử dụng các tham số LPC, MFCC với thuật giải OSALPC (linear prediction of the causal part of the autocorrelation sequence algorithm) cho mô hình GMM (Gaussian Mixture Model) trên ngữ liệu tiếng Đức (Emo-DB) đạt được độ chính xác trung bình 89% cho 7 cảm xúc.…”
Section: Các Tham Số Về Cảm Xúc Trong Tiếng Nóiunclassified
“…Speech signals have been analyzed on their own or combined with facial expression, gestures,and/or physiological signals to convey information about emotional states. More specifically, emotion recognition through speech has found increasing applications in various fields, including, but not limited to, healthcare [8], [9], [10], Human-Computer Interaction (HCI) [11], businesses [12], and entertainment [13].…”
Section: Introductionmentioning
confidence: 99%